code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position UpperCAmelCase__ = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("3.7"): raise ImportWarning( "To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition." ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( "To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n" "If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`." ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip UpperCAmelCase__ = concatenate_datasets UpperCAmelCase__ = DownloadConfig UpperCAmelCase__ = DownloadManager UpperCAmelCase__ = DownloadMode UpperCAmelCase__ = DownloadConfig UpperCAmelCase__ = DownloadMode UpperCAmelCase__ = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[str] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class __A ( unittest.TestCase ): def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) UpperCAmelCase_ = get_activation("gelu" ) self.assertTrue(torch.allclose(gelu_python(__a ) , torch_builtin(__a ) ) ) self.assertFalse(torch.allclose(gelu_python(__a ) , gelu_new(__a ) ) ) def _lowercase (self : List[str] ): UpperCAmelCase_ = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) UpperCAmelCase_ = get_activation("gelu" ) UpperCAmelCase_ = get_activation("gelu_10" ) UpperCAmelCase_ = torch_builtin(__a ) UpperCAmelCase_ = geluaa(__a ) UpperCAmelCase_ = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(__a ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def _lowercase (self : Optional[int] ): get_activation("gelu" ) get_activation("gelu_10" ) get_activation("gelu_fast" ) get_activation("gelu_new" ) get_activation("gelu_python" ) get_activation("gelu_pytorch_tanh" ) get_activation("linear" ) get_activation("mish" ) get_activation("quick_gelu" ) get_activation("relu" ) get_activation("sigmoid" ) get_activation("silu" ) get_activation("swish" ) get_activation("tanh" ) with self.assertRaises(__a ): get_activation("bogus" ) with self.assertRaises(__a ): get_activation(__a ) def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = get_activation("gelu" ) UpperCAmelCase_ = 1 UpperCAmelCase_ = get_activation("gelu" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(__a ): UpperCAmelCase_ = acta.a
1
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowercase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowercase : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowercase : set[int] = {ord(char) for char in VALID_CHARS} lowercase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str | None: _snake_case = "" _snake_case = 42 _snake_case = 42 _snake_case = 42 for keychar, cipherchar in zip(cycle(__A ) , __A ): _snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__A ) return decoded def SCREAMING_SNAKE_CASE__ ( __A ) -> list[str]: _snake_case = [] for key in product(__A , repeat=3 ): _snake_case = try_key(__A , __A ) if encoded is not None: possibles.append(__A ) return possibles def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __A = "p059_cipher.txt" ) -> int: _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = Path(__A ).parent.joinpath(__A ).read_text(encoding='utf-8' ) _snake_case = [int(__A ) for number in data.strip().split(',' )] _snake_case = filter_valid_chars(__A ) for common_word in COMMON_WORDS: _snake_case = filter_common_word(__A , __A ) if len(__A ) == 1: break _snake_case = possibles[0] return sum(ord(__A ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase : '''simple docstring''' def __init__(self : Optional[int] , UpperCamelCase : Optional[Any] , UpperCamelCase : Tuple=12 , UpperCamelCase : Tuple=7 , UpperCamelCase : List[str]=True , UpperCamelCase : Dict=True , UpperCamelCase : int=True , UpperCamelCase : str=99 , UpperCamelCase : Tuple=32 , UpperCamelCase : Dict=32 , UpperCamelCase : List[str]=2 , UpperCamelCase : int=4 , UpperCamelCase : Optional[int]=37 , UpperCamelCase : Optional[Any]=0.1 , UpperCamelCase : str=0.1 , UpperCamelCase : List[str]=512 , UpperCamelCase : Optional[int]=0.02 , UpperCamelCase : List[Any]=0 , UpperCamelCase : Union[str, Any]=None , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_input_mask lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = projection_dim lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = dropout lowercase__ = attention_dropout lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = scope lowercase__ = bos_token_id def UpperCamelCase__ (self : Optional[Any] ): '''simple docstring''' lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ = None if self.use_input_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ = input_mask.numpy() lowercase__ ,lowercase__ = input_mask.shape lowercase__ = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(UpperCamelCase ): lowercase__ = 1 lowercase__ = 0 lowercase__ = self.get_config() return config, input_ids, tf.convert_to_tensor(UpperCamelCase ) def UpperCamelCase__ (self : Any ): '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def UpperCamelCase__ (self : List[str] , UpperCamelCase : Optional[int] , UpperCamelCase : List[Any] , UpperCamelCase : List[str] ): '''simple docstring''' lowercase__ = TFBlipTextModel(config=UpperCamelCase ) lowercase__ = model(UpperCamelCase , attention_mask=UpperCamelCase , training=UpperCamelCase ) lowercase__ = model(UpperCamelCase , training=UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ ,lowercase__ ,lowercase__ = config_and_inputs lowercase__ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase (lowercase_ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ : str = (TFBlipTextModel,) if is_tf_available() else () lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : int = False lowerCAmelCase__ : Tuple = False def UpperCamelCase__ (self : Any ): '''simple docstring''' lowercase__ = BlipTextModelTester(self ) lowercase__ = ConfigTester(self , config_class=UpperCamelCase , hidden_size=37 ) def UpperCamelCase__ (self : int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ (self : str ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase ) def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' pass def UpperCamelCase__ (self : List[str] ): '''simple docstring''' pass @unittest.skip(reason='''Blip does not use inputs_embeds''' ) def UpperCamelCase__ (self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def UpperCamelCase__ (self : Dict ): '''simple docstring''' pass @unittest.skip(reason='''BlipTextModel has no base class and is not available in MODEL_MAPPING''' ) def UpperCamelCase__ (self : Optional[Any] ): '''simple docstring''' pass @slow def UpperCamelCase__ (self : List[Any] ): '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = TFBlipTextModel.from_pretrained(UpperCamelCase ) self.assertIsNotNone(UpperCamelCase ) def UpperCamelCase__ (self : Optional[Any] , UpperCamelCase : Any=True ): '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=UpperCamelCase )
2
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=32 , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=10 , SCREAMING_SNAKE_CASE=[10, 20, 30, 40] , SCREAMING_SNAKE_CASE=[1, 1, 2, 1] , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE="relu" , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=None , ) -> Dict: """simple docstring""" A : List[str] = parent A : int = batch_size A : Optional[int] = image_size A : int = num_channels A : List[Any] = embeddings_size A : str = hidden_sizes A : Any = depths A : List[str] = is_training A : Any = use_labels A : Any = hidden_act A : Tuple = num_labels A : List[str] = scope A : Union[str, Any] = len(SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A : Optional[Any] = None if self.use_labels: A : str = ids_tensor([self.batch_size] , self.num_labels ) A : List[Any] = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" A : int = RegNetModel(config=SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : List[str] = model(SCREAMING_SNAKE_CASE ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" A : str = self.num_labels A : List[Any] = RegNetForImageClassification(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : Optional[Any] = model(SCREAMING_SNAKE_CASE , labels=SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Tuple = self.prepare_config_and_inputs() A, A, A : Optional[int] = config_and_inputs A : Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class A ( __snake_case , __snake_case , unittest.TestCase ): __magic_name__ = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () __magic_name__ = ( {'''feature-extraction''': RegNetModel, '''image-classification''': RegNetForImageClassification} if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Any = RegNetModelTester(self ) A : List[str] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE , has_text_modality=SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" pass def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A, A : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A : Optional[int] = model_class(SCREAMING_SNAKE_CASE ) A : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A : Dict = [*signature.parameters.keys()] A : Optional[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A, A : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A : Tuple = model_class(config=SCREAMING_SNAKE_CASE ) for name, module in model.named_modules(): if isinstance(SCREAMING_SNAKE_CASE , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" def check_hidden_states_output(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A : Dict = model_class(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): A : Any = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) A : Any = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A : Any = self.model_tester.num_stages self.assertEqual(len(SCREAMING_SNAKE_CASE ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A, A : str = self.model_tester.prepare_config_and_inputs_for_common() A : Dict = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: A : Dict = layer_type A : Union[str, Any] = True check_hidden_states_output(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A : Dict = True check_hidden_states_output(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE ) @slow def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A : Dict = RegNetModel.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsNotNone(SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( ): '''simple docstring''' A : Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class A ( unittest.TestCase ): @cached_property def __lowerCAmelCase ( self ) -> str: """simple docstring""" return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : Optional[Any] = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(SCREAMING_SNAKE_CASE ) A : Dict = self.default_image_processor A : int = prepare_img() A : Optional[Any] = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : List[str] = model(**SCREAMING_SNAKE_CASE ) # verify the logits A : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE ) A : Tuple = torch.tensor([-0.4_180, -1.5_051, -3.4_836] ).to(SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) )
3
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
'''simple docstring''' def a_ ( lowerCamelCase : int ): lowerCAmelCase = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
4
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from typing import TYPE_CHECKING from ...utils import _LazyModule UpperCAmelCase__ = {'''tokenization_byt5''': ['''ByT5Tokenizer''']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
5
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
from __future__ import annotations from PIL import Image # Define glider example A : Union[str, Any] = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example A : Tuple = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def __lowerCAmelCase ( a__ ) -> list[list[int]]: __a = [] for i in range(len(a__ ) ): __a = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __a = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(a__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(a__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(a__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __a = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(a__ ) return next_generation def __lowerCAmelCase ( a__ , a__ ) -> list[Image.Image]: __a = [] for _ in range(a__ ): # Create output image __a = Image.new('''RGB''' , (len(cells[0] ), len(a__ )) ) __a = img.load() # Save cells to image for x in range(len(a__ ) ): for y in range(len(cells[0] ) ): __a = 255 - cells[y][x] * 255 __a = (colour, colour, colour) # Save image images.append(a__ ) __a = new_generation(a__ ) return images if __name__ == "__main__": A : Union[str, Any] = generate_images(GLIDER, 1_6) images[0].save('out.gif', save_all=True, append_images=images[1:])
6
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "google/switch-base-8": "https://huggingface.co/google/switch-base-8/blob/main/config.json", } class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 'switch_transformers' lowerCamelCase = ['past_key_values'] lowerCamelCase = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self : Dict,lowercase_ : Optional[int]=3_2_1_2_8,lowercase_ : List[Any]=7_6_8,lowercase_ : Union[str, Any]=6_4,lowercase_ : Union[str, Any]=2_0_4_8,lowercase_ : Dict=6_4,lowercase_ : List[Any]=1_2,lowercase_ : Optional[int]=3,lowercase_ : List[Any]=1_2,lowercase_ : Dict=3,lowercase_ : Any=1_2,lowercase_ : Optional[int]=8,lowercase_ : str=False,lowercase_ : Dict=0.01,lowercase_ : Optional[Any]="float32",lowercase_ : Any=False,lowercase_ : str=3_2,lowercase_ : List[Any]=1_2_8,lowercase_ : int=0.1,lowercase_ : Union[str, Any]=1E-6,lowercase_ : Dict=0.001,lowercase_ : List[Any]=0.001,lowercase_ : Dict=1.0,lowercase_ : Optional[int]="relu",lowercase_ : Dict=True,lowercase_ : Union[str, Any]=False,lowercase_ : Union[str, Any]=True,lowercase_ : List[str]=0,lowercase_ : int=1,**lowercase_ : Union[str, Any],)-> Tuple: '''simple docstring''' A__ = vocab_size A__ = d_model A__ = d_kv A__ = d_ff A__ = num_sparse_encoder_layers A__ = num_layers A__ = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry A__ = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: A__ = self.num_layers // self.num_sparse_encoder_layers else: A__ = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: A__ = self.num_decoder_layers // self.num_sparse_decoder_layers else: A__ = self.num_decoder_layers # HACK: this will create 0 sparse layers A__ = num_heads A__ = num_experts A__ = expert_capacity A__ = router_bias A__ = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F'`router_dtype` must be one of \'float32\', \'float16\' or \'bfloat16\', got {router_dtype}' ) A__ = router_dtype A__ = router_ignore_padding_tokens A__ = relative_attention_num_buckets A__ = relative_attention_max_distance A__ = dropout_rate A__ = layer_norm_epsilon A__ = initializer_factor A__ = feed_forward_proj A__ = use_cache A__ = add_router_probs A__ = router_z_loss_coef A__ = router_aux_loss_coef A__ = self.feed_forward_proj.split('-' ) A__ = act_info[-1] A__ = act_info[0] == 'gated' if len(lowercase_ ) > 1 and act_info[0] != "gated" or len(lowercase_ ) > 2: raise ValueError( F'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": A__ = 'gelu_new' super().__init__( pad_token_id=lowercase_,eos_token_id=lowercase_,is_encoder_decoder=lowercase_,**lowercase_,)
7
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class snake_case_ ( nn.Module ): '''simple docstring''' def __init__( self : Dict , _UpperCamelCase : int = 1_6 , _UpperCamelCase : int = 8_8 , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : int = 1 , _UpperCamelCase : float = 0.0 , _UpperCamelCase : int = 3_2 , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : bool = False , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : Optional[int] = None , _UpperCamelCase : str = "geglu" , _UpperCamelCase : Optional[int] = None , ) ->Any: super().__init__() snake_case_ = nn.ModuleList( [ TransformeraDModel( num_attention_heads=_UpperCamelCase , attention_head_dim=_UpperCamelCase , in_channels=_UpperCamelCase , num_layers=_UpperCamelCase , dropout=_UpperCamelCase , norm_num_groups=_UpperCamelCase , cross_attention_dim=_UpperCamelCase , attention_bias=_UpperCamelCase , sample_size=_UpperCamelCase , num_vector_embeds=_UpperCamelCase , activation_fn=_UpperCamelCase , num_embeds_ada_norm=_UpperCamelCase , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference snake_case_ = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` snake_case_ = [7_7, 2_5_7] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` snake_case_ = [1, 0] def snake_case__( self : Dict , _UpperCamelCase : str , _UpperCamelCase : str , _UpperCamelCase : Optional[Any]=None , _UpperCamelCase : List[Any]=None , _UpperCamelCase : List[str]=None , _UpperCamelCase : bool = True , ) ->Optional[Any]: snake_case_ = hidden_states snake_case_ = [] snake_case_ = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens snake_case_ = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] snake_case_ = self.transformer_index_for_condition[i] snake_case_ = self.transformers[transformer_index]( _UpperCamelCase , encoder_hidden_states=_UpperCamelCase , timestep=_UpperCamelCase , cross_attention_kwargs=_UpperCamelCase , return_dict=_UpperCamelCase , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] snake_case_ = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) snake_case_ = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=_UpperCamelCase )
8
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
import torch from diffusers import StableDiffusionPipeline __lowerCAmelCase : Dict ='path-to-your-trained-model' __lowerCAmelCase : int =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('cuda') __lowerCAmelCase : Dict ='A photo of sks dog in a bucket' __lowerCAmelCase : Dict =pipe(prompt, num_inference_steps=5_0, guidance_scale=7.5).images[0] image.save('dog-bucket.png')
9
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
from functools import lru_cache @lru_cache def lowerCAmelCase_ ( __a ) -> int: """simple docstring""" if num < 0: raise ValueError("Number should not be negative." ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
10
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version lowerCAmelCase__ = logging.getLogger(__name__) require_version('pytorch_lightning>=1.0.4') lowerCAmelCase__ = { 'base': AutoModel, 'sequence-classification': AutoModelForSequenceClassification, 'question-answering': AutoModelForQuestionAnswering, 'pretraining': AutoModelForPreTraining, 'token-classification': AutoModelForTokenClassification, 'language-modeling': AutoModelWithLMHead, 'summarization': AutoModelForSeqaSeqLM, 'translation': AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization lowerCAmelCase__ = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } lowerCAmelCase__ = sorted(arg_to_scheduler.keys()) lowerCAmelCase__ = '{' + ', '.join(arg_to_scheduler_choices) + '}' class lowerCAmelCase__ ( pl.LightningModule): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase="base" , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase , ) -> int: super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(__lowerCamelCase) _A : Tuple = 0 _A : Union[str, Any] = Path(self.hparams.output_dir) _A : str = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: _A : List[Any] = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({"num_labels": num_labels} if num_labels is not None else {}) , cache_dir=__lowerCamelCase , **__lowerCamelCase , ) else: _A : PretrainedConfig = config _A : Optional[int] = ("encoder_layerdrop", "decoder_layerdrop", "dropout", "attention_dropout") for p in extra_model_params: if getattr(self.hparams , __lowerCamelCase , __lowerCamelCase): assert hasattr(self.config , __lowerCamelCase), F"model config doesn't have a `{p}` attribute" setattr(self.config , __lowerCamelCase , getattr(self.hparams , __lowerCamelCase)) if tokenizer is None: _A : Dict = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=__lowerCamelCase , ) else: _A : PreTrainedTokenizer = tokenizer _A : List[Any] = MODEL_MODES[mode] if model is None: _A : Optional[Any] = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool(".ckpt" in self.hparams.model_name_or_path) , config=self.config , cache_dir=__lowerCamelCase , ) else: _A : Any = model def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> Union[str, Any]: _A : List[str] = self.model_type.from_pretrained(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self) -> List[str]: _A : Tuple = arg_to_scheduler[self.hparams.lr_scheduler] _A : int = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps()) _A : Union[str, Any] = {"scheduler": scheduler, "interval": "step", "frequency": 1} return scheduler def _lowerCamelCase ( self) -> Union[str, Any]: _A : Any = self.model _A : Optional[int] = ["bias", "LayerNorm.weight"] _A : int = [ { "params": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay) ], # check this named paramters "weight_decay": self.hparams.weight_decay, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], "weight_decay": 0.0, }, ] if self.hparams.adafactor: _A : Optional[int] = Adafactor( __lowerCamelCase , lr=self.hparams.learning_rate , scale_parameter=__lowerCamelCase , relative_step=__lowerCamelCase) else: _A : List[Any] = AdamW( __lowerCamelCase , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon) _A : Union[str, Any] = optimizer _A : Optional[Any] = self.get_lr_scheduler() return [optimizer], [scheduler] def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Dict: return self.validation_step(__lowerCamelCase , __lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> List[Any]: return self.validation_end(__lowerCamelCase) def _lowerCamelCase ( self) -> int: _A : int = max(1 , self.hparams.gpus) # TODO: consider num_tpu_cores _A : Union[str, Any] = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def _lowerCamelCase ( self , __lowerCamelCase) -> Dict: if stage == "test": _A : Tuple = len(self.test_dataloader().dataset) else: _A : List[str] = self.get_dataloader("train" , self.hparams.train_batch_size , shuffle=__lowerCamelCase) _A : List[str] = len(self.train_dataloader().dataset) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = False) -> Any: raise NotImplementedError("You must implement this for your task") def _lowerCamelCase ( self) -> Any: return self.train_loader def _lowerCamelCase ( self) -> List[Any]: return self.get_dataloader("dev" , self.hparams.eval_batch_size , shuffle=__lowerCamelCase) def _lowerCamelCase ( self) -> List[Any]: return self.get_dataloader("test" , self.hparams.eval_batch_size , shuffle=__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> int: return os.path.join( self.hparams.data_dir , "cached_{}_{}_{}".format( __lowerCamelCase , list(filter(__lowerCamelCase , self.hparams.model_name_or_path.split("/"))).pop() , str(self.hparams.max_seq_length) , ) , ) @pl.utilities.rank_zero_only def _lowerCamelCase ( self , __lowerCamelCase) -> None: _A : int = self.output_dir.joinpath("best_tfmr") _A : Optional[int] = self.step_count self.model.save_pretrained(__lowerCamelCase) self.tokenizer.save_pretrained(__lowerCamelCase) @staticmethod def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase) -> List[str]: parser.add_argument( "--model_name_or_path" , default=__lowerCamelCase , type=__lowerCamelCase , required=__lowerCamelCase , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--config_name" , default="" , type=__lowerCamelCase , help="Pretrained config name or path if not the same as model_name") parser.add_argument( "--tokenizer_name" , default=__lowerCamelCase , type=__lowerCamelCase , help="Pretrained tokenizer name or path if not the same as model_name" , ) parser.add_argument( "--cache_dir" , default=str(Path(__lowerCamelCase).parent / "test_run" / "cache") , type=__lowerCamelCase , help="Where do you want to store the pre-trained models downloaded from huggingface.co" , ) parser.add_argument( "--encoder_layerdrop" , type=__lowerCamelCase , help="Encoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--decoder_layerdrop" , type=__lowerCamelCase , help="Decoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--dropout" , type=__lowerCamelCase , help="Dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--attention_dropout" , type=__lowerCamelCase , help="Attention dropout probability (Optional). Goes into model.config" , ) parser.add_argument("--learning_rate" , default=5e-5 , type=__lowerCamelCase , help="The initial learning rate for Adam.") parser.add_argument( "--lr_scheduler" , default="linear" , choices=__lowerCamelCase , metavar=__lowerCamelCase , type=__lowerCamelCase , help="Learning rate scheduler" , ) parser.add_argument("--weight_decay" , default=0.0 , type=__lowerCamelCase , help="Weight decay if we apply some.") parser.add_argument("--adam_epsilon" , default=1e-8 , type=__lowerCamelCase , help="Epsilon for Adam optimizer.") parser.add_argument("--warmup_steps" , default=0 , type=__lowerCamelCase , help="Linear warmup over warmup_steps.") parser.add_argument("--num_workers" , default=4 , type=__lowerCamelCase , help="kwarg passed to DataLoader") parser.add_argument("--num_train_epochs" , dest="max_epochs" , default=3 , type=__lowerCamelCase) parser.add_argument("--train_batch_size" , default=3_2 , type=__lowerCamelCase) parser.add_argument("--eval_batch_size" , default=3_2 , type=__lowerCamelCase) parser.add_argument("--adafactor" , action="store_true") class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Optional[Any]: if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Union[str, Any]: # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(__lowerCamelCase) class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Any: _A : int = trainer.lr_schedulers[0]["scheduler"] _A : Union[str, Any] = {F"lr_group_{i}": lr for i, lr in enumerate(lr_scheduler.get_lr())} pl_module.logger.log_metrics(__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Optional[Any]: rank_zero_info("***** Validation results *****") _A : Union[str, Any] = trainer.callback_metrics # Log results for key in sorted(__lowerCamelCase): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> List[str]: rank_zero_info("***** Test results *****") _A : int = trainer.callback_metrics # Log and save results to file _A : List[Any] = os.path.join(pl_module.hparams.output_dir , "test_results.txt") with open(__lowerCamelCase , "w") as writer: for key in sorted(__lowerCamelCase): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) writer.write("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) def _UpperCAmelCase (UpperCamelCase__ : List[str] , UpperCamelCase__ : int ): # To allow all pl args uncomment the following line # parser = pl.Trainer.add_argparse_args(parser) parser.add_argument( "--output_dir" , default=str(Path(UpperCamelCase__ ).parent / "test_run" / "model_checkpoints" ) , type=UpperCamelCase__ , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument( "--fp16" , action="store_true" , help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit" , ) parser.add_argument( "--fp16_opt_level" , type=UpperCamelCase__ , default="O2" , help=( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']." "See details at https://nvidia.github.io/apex/amp.html" ) , ) parser.add_argument("--n_tpu_cores" , dest="tpu_cores" , type=UpperCamelCase__ ) parser.add_argument("--max_grad_norm" , dest="gradient_clip_val" , default=1.0 , type=UpperCamelCase__ , help="Max gradient norm" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_predict" , action="store_true" , help="Whether to run predictions on the test set." ) parser.add_argument( "--gradient_accumulation_steps" , dest="accumulate_grad_batches" , type=UpperCamelCase__ , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--seed" , type=UpperCamelCase__ , default=42 , help="random seed for initialization" ) parser.add_argument( "--data_dir" , default=str(Path(UpperCamelCase__ ).parent / "test_run" / "dummy-train-data" ) , type=UpperCamelCase__ , help="The input data dir. Should contain the training files for the CoNLL-2003 NER task." , ) def _UpperCAmelCase (UpperCamelCase__ : BaseTransformer , UpperCamelCase__ : argparse.Namespace , UpperCamelCase__ : str=None , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=[] , UpperCamelCase__ : Any=None , UpperCamelCase__ : Optional[Any]=None , **UpperCamelCase__ : Union[str, Any] , ): pl.seed_everything(args.seed ) # init model _A : str = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=UpperCamelCase__ ) # add custom checkpoints if checkpoint_callback is None: _A : Optional[Any] = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="checkpoint" , monitor="val_loss" , mode="min" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(UpperCamelCase__ ) if logging_callback is None: _A : List[Any] = LoggingCallback() _A : int = {} if args.fpaa: _A : int = 16 if args.gpus > 1: _A : str = "auto" _A : Optional[int] = "ddp" _A : List[str] = args.accumulate_grad_batches _A : str = None _A : Union[str, Any] = "auto" _A : List[str] = pl.Trainer.from_argparse_args( UpperCamelCase__ , weights_summary=UpperCamelCase__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=UpperCamelCase__ , val_check_interval=1 , num_sanity_val_steps=2 , **UpperCamelCase__ , ) if args.do_train: trainer.fit(UpperCamelCase__ ) else: print("RAG modeling tests with new set functions successfuly executed!" ) return trainer
11
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0
UpperCAmelCase_ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def lowerCamelCase__ ( ): '''simple docstring''' __lowerCamelCase = input("""Enter message: """ ) __lowerCamelCase = input("""Enter key [alphanumeric]: """ ) __lowerCamelCase = input("""Encrypt/Decrypt [e/d]: """ ) if mode.lower().startswith("""e""" ): __lowerCamelCase = """encrypt""" __lowerCamelCase = encrypt_message(A__ , A__ ) elif mode.lower().startswith("""d""" ): __lowerCamelCase = """decrypt""" __lowerCamelCase = decrypt_message(A__ , A__ ) print(f'\n{mode.title()}ed message:' ) print(A__ ) def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' return translate_message(A__ , A__ , """encrypt""" ) def lowerCamelCase__ ( A__ : str , A__ : str ): '''simple docstring''' return translate_message(A__ , A__ , """decrypt""" ) def lowerCamelCase__ ( A__ : str , A__ : str , A__ : str ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = 0 __lowerCamelCase = key.upper() for symbol in message: __lowerCamelCase = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(A__ ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(A__ ): __lowerCamelCase = 0 else: translated.append(A__ ) return "".join(A__ ) if __name__ == "__main__": main()
12
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __lowercase ( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Optional[int] = BertJapaneseTokenizer _UpperCAmelCase : Any = False _UpperCAmelCase : Dict = True def _SCREAMING_SNAKE_CASE ( self : Optional[int]): super().setUp() SCREAMING_SNAKE_CASE_: Dict = [ "[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは", "世界", "##世界", "、", "##、", "。", "##。", ] SCREAMING_SNAKE_CASE_: Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : Optional[int]): SCREAMING_SNAKE_CASE_: Optional[int] = "こんにちは、世界。 \nこんばんは、世界。" SCREAMING_SNAKE_CASE_: Dict = "こんにちは 、 世界 。 こんばんは 、 世界 。" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase__ : List[str]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = self.get_input_output_texts(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = tokenizer.decode(lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__) return text, ids def _SCREAMING_SNAKE_CASE ( self : List[Any]): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : Dict): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : Optional[int]): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: int = self.tokenizer_class(self.vocab_file) SCREAMING_SNAKE_CASE_: List[Any] = tokenizer.tokenize("こんにちは、世界。\nこんばんは、世界。") self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [3, 12, 10, 14, 4, 9, 12, 10, 14]) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_: int = self.tokenizer_class(self.vocab_file , word_tokenizer_type="mecab") self.assertIsNotNone(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = "こんにちは、世界。\nこんばんは、世界。" SCREAMING_SNAKE_CASE_: Optional[int] = tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [3, 12, 10, 14, 4, 9, 12, 10, 14]) SCREAMING_SNAKE_CASE_: int = os.path.join(self.tmpdirname , "tokenizer.bin") with open(lowerCAmelCase__ , "wb") as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__) with open(lowerCAmelCase__ , "rb") as handle: SCREAMING_SNAKE_CASE_: List[Any] = pickle.load(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = tokenizer_new.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Union[str, Any] = MecabTokenizer(mecab_dic="ipadic") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): try: SCREAMING_SNAKE_CASE_: List[str] = MecabTokenizer(mecab_dic="unidic_lite") except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : str): try: SCREAMING_SNAKE_CASE_: Tuple = MecabTokenizer(mecab_dic="unidic") except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Tuple = MecabTokenizer(do_lower_case=lowerCAmelCase__ , mecab_dic="ipadic") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップルストア", "で", "iphone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : int): try: SCREAMING_SNAKE_CASE_: int = MecabTokenizer( do_lower_case=lowerCAmelCase__ , normalize_text=lowerCAmelCase__ , mecab_option="-d /usr/local/lib/mecab/dic/jumandic") except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "\u3000", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Dict = MecabTokenizer(normalize_text=lowerCAmelCase__ , mecab_dic="ipadic") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", " ", "。"] , ) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: Any = self.tokenizer_class(self.vocab_file , word_tokenizer_type="sudachi") self.assertIsNotNone(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = "こんにちは、世界。\nこんばんは、世界。" SCREAMING_SNAKE_CASE_: Any = tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [3, 12, 10, 14, 4, 9, 12, 10, 14]) SCREAMING_SNAKE_CASE_: Tuple = os.path.join(self.tmpdirname , "tokenizer.bin") with open(lowerCAmelCase__ , "wb") as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__) with open(lowerCAmelCase__ , "rb") as handle: SCREAMING_SNAKE_CASE_: str = pickle.load(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = tokenizer_new.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_: Optional[int] = SudachiTokenizer(sudachi_dict_type="core") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_: Union[str, Any] = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="A") self.assertListEqual(tokenizer.tokenize("外国人参政権") , ["外国", "人", "参政", "権"]) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Tuple = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="B") self.assertListEqual(tokenizer.tokenize("外国人参政権") , ["外国人", "参政権"]) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_: Union[str, Any] = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="C") self.assertListEqual(tokenizer.tokenize("外国人参政権") , ["外国人参政権"]) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: Optional[Any] = SudachiTokenizer(do_lower_case=lowerCAmelCase__ , sudachi_dict_type="core") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , [" ", "\t", "アップル", "ストア", "で", "iphone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_: str = SudachiTokenizer(normalize_text=lowerCAmelCase__ , sudachi_dict_type="core") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", "\u3000", "。", " ", " "] , ) @require_sudachi def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_: Optional[Any] = SudachiTokenizer(trim_whitespace=lowerCAmelCase__ , sudachi_dict_type="core") self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Tuple = self.tokenizer_class(self.vocab_file , word_tokenizer_type="jumanpp") self.assertIsNotNone(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = "こんにちは、世界。\nこんばんは、世界。" SCREAMING_SNAKE_CASE_: Dict = tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [3, 12, 10, 14, 4, 9, 12, 10, 14]) SCREAMING_SNAKE_CASE_: Optional[int] = os.path.join(self.tmpdirname , "tokenizer.bin") with open(lowerCAmelCase__ , "wb") as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__) with open(lowerCAmelCase__ , "rb") as handle: SCREAMING_SNAKE_CASE_: Dict = pickle.load(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = tokenizer_new.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_: Union[str, Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: List[str] = JumanppTokenizer(do_lower_case=lowerCAmelCase__) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iphone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Dict = JumanppTokenizer(normalize_text=lowerCAmelCase__) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["ア", "ッ", "フ", "゚", "ル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_: str = JumanppTokenizer(trim_whitespace=lowerCAmelCase__) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 ") , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "。"] , ) @require_jumanpp def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Union[str, Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("ありがとうございますm(_ _)m見つけるのが大変です。") , ["ありがとう", "ございます", "m(_ _)m", "見つける", "の", "が", "大変です", "。"] , ) def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_: Optional[Any] = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは"] SCREAMING_SNAKE_CASE_: Optional[int] = {} for i, token in enumerate(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: Optional[Any] = i SCREAMING_SNAKE_CASE_: Union[str, Any] = WordpieceTokenizer(vocab=lowerCAmelCase__ , unk_token="[UNK]") self.assertListEqual(tokenizer.tokenize("") , []) self.assertListEqual(tokenizer.tokenize("こんにちは") , ["こんにちは"]) self.assertListEqual(tokenizer.tokenize("こんばんは") , ["こん", "##ばんは"]) self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは") , ["こん", "##ばんは", "[UNK]", "こんにちは"]) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: str = BertJapaneseTokenizer.from_pretrained("nlp-waseda/roberta-base-japanese-with-auto-jumanpp") SCREAMING_SNAKE_CASE_: List[Any] = tokenizer.subword_tokenizer SCREAMING_SNAKE_CASE_: List[Any] = subword_tokenizer.tokenize("国境 の 長い トンネル を 抜ける と 雪国 であった 。") self.assertListEqual(lowerCAmelCase__ , ["▁国境", "▁の", "▁長い", "▁トンネル", "▁を", "▁抜ける", "▁と", "▁雪", "国", "▁であった", "▁。"]) SCREAMING_SNAKE_CASE_: str = subword_tokenizer.tokenize("こんばんは こんばん にち は こんにちは") self.assertListEqual(lowerCAmelCase__ , ["▁こん", "ばん", "は", "▁こん", "ばん", "▁に", "ち", "▁は", "▁こんにちは"]) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Dict = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese") SCREAMING_SNAKE_CASE_: Tuple = tokenizer.encode("ありがとう。" , add_special_tokens=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = tokenizer.encode("どういたしまして。" , add_special_tokens=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Optional[Any] = BertJapaneseTokenizer _UpperCAmelCase : List[str] = False def _SCREAMING_SNAKE_CASE ( self : List[Any]): super().setUp() SCREAMING_SNAKE_CASE_: Any = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] SCREAMING_SNAKE_CASE_: int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self : List[str] , **lowerCAmelCase__ : Union[str, Any]): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="character" , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : Union[str, Any]): SCREAMING_SNAKE_CASE_: List[Any] = "こんにちは、世界。 \nこんばんは、世界。" SCREAMING_SNAKE_CASE_: Union[str, Any] = "こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self : Tuple): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : int): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : Any): pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_: Dict = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="character") SCREAMING_SNAKE_CASE_: Optional[int] = tokenizer.tokenize("こんにちは、世界。 \nこんばんは、世界。") self.assertListEqual( lowerCAmelCase__ , ["こ", "ん", "に", "ち", "は", "、", "世", "界", "。", "こ", "ん", "ば", "ん", "は", "、", "世", "界", "。"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12]) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Optional[Any] = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] SCREAMING_SNAKE_CASE_: List[Any] = {} for i, token in enumerate(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: Dict = i SCREAMING_SNAKE_CASE_: Optional[int] = CharacterTokenizer(vocab=lowerCAmelCase__ , unk_token="[UNK]") self.assertListEqual(tokenizer.tokenize("") , []) self.assertListEqual(tokenizer.tokenize("こんにちは") , ["こ", "ん", "に", "ち", "は"]) self.assertListEqual(tokenizer.tokenize("こんにちほ") , ["こ", "ん", "に", "ち", "[UNK]"]) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: str = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese-char") SCREAMING_SNAKE_CASE_: Optional[int] = tokenizer.encode("ありがとう。" , add_special_tokens=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = tokenizer.encode("どういたしまして。" , add_special_tokens=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __lowercase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_: List[str] = "cl-tohoku/bert-base-japanese" SCREAMING_SNAKE_CASE_: List[str] = AutoTokenizer.from_pretrained(lowerCAmelCase__) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__) class __lowercase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Union[str, Any] = "cl-tohoku/bert-base-japanese" with self.assertLogs("transformers" , level="WARNING") as cm: BertTokenizer.from_pretrained(lowerCAmelCase__) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from.")) SCREAMING_SNAKE_CASE_: Union[str, Any] = "bert-base-cased" with self.assertLogs("transformers" , level="WARNING") as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase__) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from."))
13
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' @slow @require_torch def SCREAMING_SNAKE_CASE ( self : Tuple) ->List[Any]: '''simple docstring''' A__ = EncoderDecoderModel.from_encoder_decoder_pretrained('''prajjwal1/bert-tiny''' , '''prajjwal1/bert-tiny''') A__ = BertTokenizer.from_pretrained('''bert-base-uncased''') A__ = bertabert.config.encoder.vocab_size A__ = tokenizer.sep_token_id A__ = tokenizer.cls_token_id A__ = 128 A__ = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''train[:1%]''') A__ = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''validation[:1%]''') A__ = train_dataset.select(range(32)) A__ = val_dataset.select(range(16)) A__ = 4 def _map_to_encoder_decoder_inputs(UpperCAmelCase__ : List[str]): # Tokenizer will automatically set [BOS] <text> [EOS] A__ = tokenizer(batch['''article'''] , padding='''max_length''' , truncation=UpperCAmelCase__ , max_length=512) A__ = tokenizer(batch['''highlights'''] , padding='''max_length''' , truncation=UpperCAmelCase__ , max_length=128) A__ = inputs.input_ids A__ = inputs.attention_mask A__ = outputs.input_ids A__ = outputs.input_ids.copy() A__ = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['''labels'''] ] A__ = outputs.attention_mask assert all(len(UpperCAmelCase__) == 512 for x in inputs.input_ids) assert all(len(UpperCAmelCase__) == 128 for x in outputs.input_ids) return batch def _compute_metrics(UpperCAmelCase__ : Tuple): A__ = pred.label_ids A__ = pred.predictions # all unnecessary tokens are removed A__ = tokenizer.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__) A__ = tokenizer.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__) A__ = sum([int(pred_str[i] == label_str[i]) for i in range(len(UpperCAmelCase__))]) / len(UpperCAmelCase__) return {"accuracy": accuracy} # map train dataset A__ = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCAmelCase__ , batch_size=UpperCAmelCase__ , remove_columns=['''article''', '''highlights'''] , ) train_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) # same for validation dataset A__ = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCAmelCase__ , batch_size=UpperCAmelCase__ , remove_columns=['''article''', '''highlights'''] , ) val_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) A__ = self.get_auto_remove_tmp_dir() A__ = SeqaSeqTrainingArguments( output_dir=UpperCAmelCase__ , per_device_train_batch_size=UpperCAmelCase__ , per_device_eval_batch_size=UpperCAmelCase__ , predict_with_generate=UpperCAmelCase__ , evaluation_strategy='''steps''' , do_train=UpperCAmelCase__ , do_eval=UpperCAmelCase__ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer A__ = SeqaSeqTrainer( model=UpperCAmelCase__ , args=UpperCAmelCase__ , compute_metrics=_compute_metrics , train_dataset=UpperCAmelCase__ , eval_dataset=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , ) # start training trainer.train()
14
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowercase : Optional[Any] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") lowercase : Tuple = parser.parse_args() lowercase : Optional[int] = "cpu" lowercase : Optional[Any] = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowercase : Optional[int] = "path-to-your-trained-model" lowercase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowercase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowercase : Dict = pipe.to(device) # to channels last lowercase : Optional[Any] = pipe.unet.to(memory_format=torch.channels_last) lowercase : int = pipe.vae.to(memory_format=torch.channels_last) lowercase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowercase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowercase : Any = torch.randn(2, 4, 64, 64) lowercase : Optional[int] = torch.rand(1) * 999 lowercase : Optional[Any] = torch.randn(2, 77, 768) lowercase : Optional[Any] = (sample, timestep, encoder_hidden_status) try: lowercase : List[Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowercase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Optional[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowercase : Tuple = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowercase : List[str] = 666 lowercase : Tuple = torch.Generator(device).manual_seed(seed) lowercase : Union[str, Any] = {"generator": generator} if args.steps is not None: lowercase : Dict = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowercase : List[str] = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
42
0
def UpperCAmelCase ( a_ = 1_0_0_0 ) -> int: """simple docstring""" return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
15
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCAmelCase ( _lowerCamelCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return 0.0 def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[int | float, int | float]: _snake_case = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.abs(np.fft.fft(__A ) ) _snake_case = 20 * np.logaa(__A ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds _snake_case = get_bounds(__A , __A ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__A ) plt.show() def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.angle(np.fft.fft(__A ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__A , -2 * pi ) ) plt.show()
42
0
"""simple docstring""" import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> List[str]: # Initialise PyTorch model lowercase__ : List[Any] = RemBertConfig.from_json_file(__lowerCamelCase ) print('''Building PyTorch model from configuration: {}'''.format(str(__lowerCamelCase ) ) ) lowercase__ : Dict = RemBertModel(__lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_rembert(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print('''Save PyTorch model to {}'''.format(__lowerCamelCase ) ) torch.save(model.state_dict() , __lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--rembert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained RemBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowerCAmelCase_ = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
16
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class __UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = vocab_size _snake_case = d_embed _snake_case = d_proj _snake_case = cutoffs + [vocab_size] _snake_case = [0] + self.cutoffs _snake_case = div_val _snake_case = self.cutoffs[0] _snake_case = len(self.cutoffs ) - 1 _snake_case = self.shortlist_size + self.n_clusters _snake_case = keep_order _snake_case = [] _snake_case = [] def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if self.n_clusters > 0: _snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_weight' ) _snake_case = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: _snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' , ) self.out_projs.append(lowerCAmelCase_ ) else: self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case = self.d_embed // (self.div_val**i) _snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' ) self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = x if proj is not None: _snake_case = tf.einsum('ibd,ed->ibe' , lowerCAmelCase_ , lowerCAmelCase_ ) return tf.einsum('ibd,nd->ibn' , lowerCAmelCase_ , lowerCAmelCase_ ) + b @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = tf.range(lp_size[0] , dtype=target.dtype ) _snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False ): """simple docstring""" _snake_case = 0 if self.n_clusters == 0: _snake_case = self._logit(lowerCAmelCase_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: _snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase_ , logits=lowerCAmelCase_ ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ , axis=-1 ) else: _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = [] _snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: _snake_case = (target >= l_idx) & (target < r_idx) _snake_case = tf.where(lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) - l_idx if self.div_val == 1: _snake_case = self.out_layers[0][0][l_idx:r_idx] _snake_case = self.out_layers[0][1][l_idx:r_idx] else: _snake_case = self.out_layers[i][0] _snake_case = self.out_layers[i][1] if i == 0: _snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) _snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[0] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) else: _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[i] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) _snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster _snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase_ ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase_ , -cur_logprob , shape_list(lowerCAmelCase_ ) ) _snake_case = tf.concat(lowerCAmelCase_ , axis=-1 ) if target is not None: if return_mean: _snake_case = tf.reduce_mean(lowerCAmelCase_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase_ , name=self.name , aggregation='mean' if return_mean else '' ) return out
42
0
"""simple docstring""" def _A ( UpperCamelCase_ : list[list[int]], UpperCamelCase_ : int, UpperCamelCase_ : int, UpperCamelCase_ : list[int]) -> bool: '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path) def _A ( UpperCamelCase_ : list[list[int]], UpperCamelCase_ : list[int], UpperCamelCase_ : int) -> bool: '''simple docstring''' if curr_ind == len(UpperCamelCase_): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0, len(UpperCamelCase_)): if valid_connection(UpperCamelCase_, UpperCamelCase_, UpperCamelCase_, UpperCamelCase_): # Insert current vertex into path as next transition __lowercase = next_ver # Validate created path if util_hamilton_cycle(UpperCamelCase_, UpperCamelCase_, curr_ind + 1): return True # Backtrack __lowercase = -1 return False def _A ( UpperCamelCase_ : list[list[int]], UpperCamelCase_ : int = 0) -> list[int]: '''simple docstring''' __lowercase = [-1] * (len(UpperCamelCase_) + 1) # initialize start and end of path with starting index __lowercase = __lowercase = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCamelCase_, UpperCamelCase_, 1) else []
17
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
0
__lowerCamelCase : Tuple = 2_56 # Modulus to hash a string __lowerCamelCase : int = 1_00_00_03 def _snake_case ( lowerCAmelCase : str , lowerCAmelCase : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = len(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = len(lowerCAmelCase ) if p_len > t_len: return False SCREAMING_SNAKE_CASE_ : str = 0 SCREAMING_SNAKE_CASE_ : Dict = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = 1 # Calculating the hash of pattern and substring of text for i in range(lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : List[str] = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus SCREAMING_SNAKE_CASE_ : List[Any] = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue SCREAMING_SNAKE_CASE_ : List[str] = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def _snake_case ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = "abc1abc12" SCREAMING_SNAKE_CASE_ : Optional[int] = "alskfjaldsabc1abc1abc12k23adsfabcabc" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "alskfjaldsk23adsfabcabc" assert rabin_karp(lowerCAmelCase , lowerCAmelCase ) and not rabin_karp(lowerCAmelCase , lowerCAmelCase ) # Test 2) SCREAMING_SNAKE_CASE_ : List[Any] = "ABABX" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "ABABZABABYABABX" assert rabin_karp(lowerCAmelCase , lowerCAmelCase ) # Test 3) SCREAMING_SNAKE_CASE_ : Dict = "AAAB" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "ABAAAAAB" assert rabin_karp(lowerCAmelCase , lowerCAmelCase ) # Test 4) SCREAMING_SNAKE_CASE_ : Tuple = "abcdabcy" SCREAMING_SNAKE_CASE_ : Optional[int] = "abcxabcdabxabcdabcdabcy" assert rabin_karp(lowerCAmelCase , lowerCAmelCase ) # Test 5) SCREAMING_SNAKE_CASE_ : List[Any] = "Lü" SCREAMING_SNAKE_CASE_ : List[Any] = "Lüsai" assert rabin_karp(lowerCAmelCase , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = "Lue" assert not rabin_karp(lowerCAmelCase , lowerCAmelCase ) print("Success." ) if __name__ == "__main__": test_rabin_karp()
18
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def lowerCamelCase_ ( lowerCamelCase__ ): return {key.lstrip("-" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def lowerCamelCase_ ( ): lowerCamelCase_ = ArgumentParser( "HuggingFace Datasets CLI tool" , usage="datasets-cli <command> [<args>]" , allow_abbrev=lowerCamelCase__ ) lowerCamelCase_ = parser.add_subparsers(help="datasets-cli command helpers" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(lowerCamelCase__ ) EnvironmentCommand.register_subcommand(lowerCamelCase__ ) TestCommand.register_subcommand(lowerCamelCase__ ) RunBeamCommand.register_subcommand(lowerCamelCase__ ) DummyDataCommand.register_subcommand(lowerCamelCase__ ) # Parse args lowerCamelCase_ , lowerCamelCase_ = parser.parse_known_args() if not hasattr(lowerCamelCase__ , "func" ): parser.print_help() exit(1 ) lowerCamelCase_ = parse_unknown_args(lowerCamelCase__ ) # Run lowerCamelCase_ = args.func(lowerCamelCase__ , **lowerCamelCase__ ) service.run() if __name__ == "__main__": main()
19
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = 1 _snake_case = 2 while i * i <= n: _snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _snake_case = 1 _snake_case = 1 while True: i += 1 t_num += i if count_divisors(__A ) > 500: break return t_num if __name__ == "__main__": print(solution())
42
0
def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> int: lowercase , lowercase : Union[str, Any] = len(SCREAMING_SNAKE_CASE__ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase : str = 0 count += depth_first_search(SCREAMING_SNAKE_CASE__ , row + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) count += depth_first_search(SCREAMING_SNAKE_CASE__ , row - 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) count += depth_first_search(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , col + 1 , SCREAMING_SNAKE_CASE__ ) count += depth_first_search(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , col - 1 , SCREAMING_SNAKE_CASE__ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
20
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __UpperCAmelCase ( _lowerCamelCase ): @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _snake_case = self.get_env() _snake_case = '1' _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import AutoModel\n ' _snake_case = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
42
0
from __future__ import annotations class _lowerCamelCase: def __init__( self, lowerCamelCase=None) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = data _lowercase : List[str] = None def __repr__( self) -> Optional[int]: """simple docstring""" _lowercase : Tuple = [] _lowercase : Dict = self while temp: string_rep.append(F'''{temp.data}''') _lowercase : int = temp.next return "->".join(lowerCamelCase) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: if not elements_list: raise Exception('The Elements List is empty' ) _lowercase : List[str] = Node(elements_list[0] ) for i in range(1 , len(lowerCamelCase_ ) ): _lowercase : Optional[int] = Node(elements_list[i] ) _lowercase : str = current.next return head def UpperCamelCase_( lowerCamelCase_ ) -> None: if head_node is not None and isinstance(lowerCamelCase_ , lowerCamelCase_ ): print_reverse(head_node.next ) print(head_node.data ) def UpperCamelCase_( ) -> Dict: from doctest import testmod testmod() _lowercase : str = make_linked_list([14, 52, 14, 12, 43] ) print('Linked List:' ) print(lowerCamelCase_ ) print('Elements in Reverse:' ) print_reverse(lowerCamelCase_ ) if __name__ == "__main__": main()
21
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCAmelCase_ ).to(lowerCAmelCase_ ) _snake_case = AutoTokenizer.from_pretrained('google/mt5-small' ) _snake_case = tokenizer('Hello there' , return_tensors='pt' ).input_ids _snake_case = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _snake_case = model(input_ids.to(lowerCAmelCase_ ) , labels=labels.to(lowerCAmelCase_ ) ).loss _snake_case = -(labels.shape[-1] * loss.item()) _snake_case = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
42
0
'''simple docstring''' def UpperCAmelCase_ ( __lowercase : int , __lowercase : Optional[int] ) -> str: '''simple docstring''' _UpperCAmelCase = 0 _UpperCAmelCase = len(__lowercase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None _UpperCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowercase ): return None _UpperCAmelCase = sorted_collection[point] if current_item == item: return point else: if point < left: _UpperCAmelCase = left _UpperCAmelCase = point elif point > right: _UpperCAmelCase = right _UpperCAmelCase = point else: if item < current_item: _UpperCAmelCase = point - 1 else: _UpperCAmelCase = point + 1 return None def UpperCAmelCase_ ( __lowercase : Any , __lowercase : List[Any] , __lowercase : Dict , __lowercase : List[str] ) -> Union[str, Any]: '''simple docstring''' if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None _UpperCAmelCase = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__lowercase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__lowercase , __lowercase , __lowercase , __lowercase ) elif point > right: return interpolation_search_by_recursion(__lowercase , __lowercase , __lowercase , __lowercase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __lowercase , __lowercase , __lowercase , point - 1 ) else: return interpolation_search_by_recursion( __lowercase , __lowercase , point + 1 , __lowercase ) def UpperCAmelCase_ ( __lowercase : int ) -> Tuple: '''simple docstring''' if collection != sorted(__lowercase ): raise ValueError("Collection must be ascending sorted" ) return True if __name__ == "__main__": import sys __SCREAMING_SNAKE_CASE :Optional[int] = 0 if debug == 1: __SCREAMING_SNAKE_CASE :Union[str, Any] = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit('''Sequence must be ascending sorted to apply interpolation search''') __SCREAMING_SNAKE_CASE :int = 67 __SCREAMING_SNAKE_CASE :Union[str, Any] = interpolation_search(collection, target) if result is not None: print(F"{target} found at positions: {result}") else: print('''Not found''')
22
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[str] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
'''simple docstring''' from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : bool = False ) -> list[float]: if radian_mode: return [magnitude * cos(_lowerCAmelCase ), magnitude * sin(_lowerCAmelCase )] return [magnitude * cos(radians(_lowerCAmelCase ) ), magnitude * sin(radians(_lowerCAmelCase ) )] def snake_case_ ( _lowerCAmelCase : NDArray[floataa] , _lowerCAmelCase : NDArray[floataa] , _lowerCAmelCase : float = 10**-1 ) -> bool: UpperCAmelCase : NDArray[floataa] = cross(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : float = sum(_lowerCAmelCase ) return abs(_lowerCAmelCase ) < eps if __name__ == "__main__": # Test to check if it works UpperCamelCase__: List[Any] = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) UpperCamelCase__: NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg UpperCamelCase__: Optional[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) UpperCamelCase__: int = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg UpperCamelCase__: List[Any] = array([[0, -2000], [0, -1200], [0, 15600], [0, -12400]]) UpperCamelCase__: List[str] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
23
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowercase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowercase : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowercase : set[int] = {ord(char) for char in VALID_CHARS} lowercase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str | None: _snake_case = "" _snake_case = 42 _snake_case = 42 _snake_case = 42 for keychar, cipherchar in zip(cycle(__A ) , __A ): _snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__A ) return decoded def SCREAMING_SNAKE_CASE__ ( __A ) -> list[str]: _snake_case = [] for key in product(__A , repeat=3 ): _snake_case = try_key(__A , __A ) if encoded is not None: possibles.append(__A ) return possibles def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __A = "p059_cipher.txt" ) -> int: _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = Path(__A ).parent.joinpath(__A ).read_text(encoding='utf-8' ) _snake_case = [int(__A ) for number in data.strip().split(',' )] _snake_case = filter_valid_chars(__A ) for common_word in COMMON_WORDS: _snake_case = filter_common_word(__A , __A ) if len(__A ) == 1: break _snake_case = possibles[0] return sum(ord(__A ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class SCREAMING_SNAKE_CASE__ : A_ : int A_ : TreeNode | None = None A_ : TreeNode | None = None snake_case_ = namedtuple('CoinsDistribResult', 'moves excess') def lowerCamelCase__ ( snake_case_ : TreeNode | None ) -> int: if root is None: return 0 # Validation def count_nodes(snake_case_ : TreeNode | None ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(snake_case_ : TreeNode | None ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(snake_case_ ) != count_coins(snake_case_ ): raise ValueError('''The nodes number should be same as the number of coins''' ) # Main calculation def get_distrib(snake_case_ : TreeNode | None ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) __snake_case , __snake_case = get_distrib(node.left ) __snake_case , __snake_case = get_distrib(node.right ) __snake_case = 1 - left_distrib_excess __snake_case = 1 - right_distrib_excess __snake_case = ( left_distrib_moves + right_distrib_moves + abs(snake_case_ ) + abs(snake_case_ ) ) __snake_case = node.data - coins_to_left - coins_to_right return CoinsDistribResult(snake_case_ , snake_case_ ) return get_distrib(snake_case_ )[0] if __name__ == "__main__": import doctest doctest.testmod()
24
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCAmelCase_ : """simple docstring""" def __init__(self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=13 , SCREAMING_SNAKE_CASE__=30 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=32 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=4 , SCREAMING_SNAKE_CASE__=37 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=0.02 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=2 , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = parent SCREAMING_SNAKE_CASE__ : List[Any] = batch_size SCREAMING_SNAKE_CASE__ : Any = image_size SCREAMING_SNAKE_CASE__ : Tuple = patch_size SCREAMING_SNAKE_CASE__ : Tuple = num_channels SCREAMING_SNAKE_CASE__ : Union[str, Any] = is_training SCREAMING_SNAKE_CASE__ : List[str] = use_labels SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_size SCREAMING_SNAKE_CASE__ : Any = num_hidden_layers SCREAMING_SNAKE_CASE__ : int = num_attention_heads SCREAMING_SNAKE_CASE__ : Dict = intermediate_size SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_act SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_dropout_prob SCREAMING_SNAKE_CASE__ : List[Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : int = type_sequence_label_size SCREAMING_SNAKE_CASE__ : List[str] = initializer_range SCREAMING_SNAKE_CASE__ : Optional[Any] = scope SCREAMING_SNAKE_CASE__ : Union[str, Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) SCREAMING_SNAKE_CASE__ : Union[str, Any] = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE__ : List[Any] = num_patches + 2 def __magic_name__ (self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ : List[str] = None if self.use_labels: SCREAMING_SNAKE_CASE__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ : int = self.get_config() return config, pixel_values, labels def __magic_name__ (self ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = DeiTModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() SCREAMING_SNAKE_CASE__ : int = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = DeiTForMaskedImageModeling(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() SCREAMING_SNAKE_CASE__ : List[str] = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images SCREAMING_SNAKE_CASE__ : int = 1 SCREAMING_SNAKE_CASE__ : List[str] = DeiTForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() SCREAMING_SNAKE_CASE__ : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ : Optional[int] = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = self.type_sequence_label_size SCREAMING_SNAKE_CASE__ : List[Any] = DeiTForImageClassification(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() SCREAMING_SNAKE_CASE__ : int = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE__ : Tuple = 1 SCREAMING_SNAKE_CASE__ : Optional[int] = DeiTForImageClassification(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() SCREAMING_SNAKE_CASE__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ : str = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ (self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) : List[Any] = config_and_inputs SCREAMING_SNAKE_CASE__ : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ (a__ , a__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Any = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) __UpperCamelCase : List[str] = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) __UpperCamelCase : List[Any] = False __UpperCamelCase : Union[str, Any] = False __UpperCamelCase : Tuple = False def __magic_name__ (self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = DeiTModelTester(self ) SCREAMING_SNAKE_CASE__ : List[Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ , hidden_size=37 ) def __magic_name__ (self ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def __magic_name__ (self ) -> List[Any]: """simple docstring""" pass def __magic_name__ (self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : Any = model_class(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE__ : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE__ , nn.Linear ) ) def __magic_name__ (self ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : int = model_class(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ : Optional[int] = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = super()._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __magic_name__ (self ) -> List[Any]: """simple docstring""" if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ : List[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(SCREAMING_SNAKE_CASE__ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue SCREAMING_SNAKE_CASE__ : Union[str, Any] = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.train() SCREAMING_SNAKE_CASE__ : Any = self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = model(**SCREAMING_SNAKE_CASE__ ).loss loss.backward() def __magic_name__ (self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE__ : List[Any] = False SCREAMING_SNAKE_CASE__ : Optional[Any] = True for model_class in self.all_model_classes: if model_class in get_values(SCREAMING_SNAKE_CASE__ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue SCREAMING_SNAKE_CASE__ : Optional[int] = model_class(SCREAMING_SNAKE_CASE__ ) model.gradient_checkpointing_enable() model.to(SCREAMING_SNAKE_CASE__ ) model.train() SCREAMING_SNAKE_CASE__ : int = self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = model(**SCREAMING_SNAKE_CASE__ ).loss loss.backward() def __magic_name__ (self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ : int = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(SCREAMING_SNAKE_CASE__ ), *get_values(SCREAMING_SNAKE_CASE__ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type['title']}''' ): SCREAMING_SNAKE_CASE__ : Dict = problem_type["""title"""] SCREAMING_SNAKE_CASE__ : List[Any] = problem_type["""num_labels"""] SCREAMING_SNAKE_CASE__ : Dict = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.train() SCREAMING_SNAKE_CASE__ : List[Any] = self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) if problem_type["num_labels"] > 1: SCREAMING_SNAKE_CASE__ : List[str] = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=SCREAMING_SNAKE_CASE__ ) as warning_list: SCREAMING_SNAKE_CASE__ : List[Any] = model(**SCREAMING_SNAKE_CASE__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def __magic_name__ (self ) -> List[Any]: """simple docstring""" for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ : Tuple = DeiTModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def lowercase_ ( ): SCREAMING_SNAKE_CASE__ : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowerCAmelCase_ (unittest.TestCase ): """simple docstring""" @cached_property def __magic_name__ (self ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def __magic_name__ (self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = self.default_image_processor SCREAMING_SNAKE_CASE__ : str = prepare_img() SCREAMING_SNAKE_CASE__ : List[Any] = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Union[str, Any] = model(**SCREAMING_SNAKE_CASE__ ) # verify the logits SCREAMING_SNAKE_CASE__ : Tuple = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[str] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def __magic_name__ (self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) SCREAMING_SNAKE_CASE__ : Any = self.default_image_processor SCREAMING_SNAKE_CASE__ : List[Any] = prepare_img() SCREAMING_SNAKE_CASE__ : Optional[Any] = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ : Optional[int] = inputs.pixel_values.to(SCREAMING_SNAKE_CASE__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Tuple = model(SCREAMING_SNAKE_CASE__ )
25
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( UpperCamelCase__ ): _a = (UnCLIPScheduler,) def a__ ( self , **_a ) -> Any: _A : Dict = { """num_train_timesteps""": 1000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**_a ) return config def a__ ( self ) -> Union[str, Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=_a ) def a__ ( self ) -> str: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=_a ) def a__ ( self ) -> List[str]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_a ) def a__ ( self ) -> Optional[Any]: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=_a ) def a__ ( self ) -> str: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=_a ) def a__ ( self ) -> Tuple: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=_a , prev_timestep=_a ) def a__ ( self ) -> str: _A : str = self.scheduler_classes[0] _A : str = self.get_scheduler_config(variance_type="""fixed_small_log""" ) _A : Optional[Any] = scheduler_class(**_a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.00_00e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0549625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.9994987 ) ) < 1e-5 def a__ ( self ) -> int: _A : Dict = self.scheduler_classes[0] _A : List[str] = self.get_scheduler_config(variance_type="""learned_range""" ) _A : Optional[Any] = scheduler_class(**_a ) _A : str = 0.5 assert scheduler._get_variance(1 , predicted_variance=_a ) - -10.1712790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=_a ) - -5.7998052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=_a ) - -0.0010011 < 1e-5 def a__ ( self ) -> Any: _A : Any = self.scheduler_classes[0] _A : Tuple = self.get_scheduler_config() _A : Optional[Any] = scheduler_class(**_a ) _A : Optional[int] = scheduler.timesteps _A : Union[str, Any] = self.dummy_model() _A : Any = self.dummy_sample_deter _A : List[str] = torch.manual_seed(0 ) for i, t in enumerate(_a ): # 1. predict noise residual _A : Dict = model(_a , _a ) # 2. predict previous mean of sample x_t-1 _A : List[Any] = scheduler.step(_a , _a , _a , generator=_a ).prev_sample _A : Any = pred_prev_sample _A : int = torch.sum(torch.abs(_a ) ) _A : Optional[int] = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 252.2682495 ) < 1e-2 assert abs(result_mean.item() - 0.3284743 ) < 1e-3 def a__ ( self ) -> Any: _A : Dict = self.scheduler_classes[0] _A : Optional[int] = self.get_scheduler_config() _A : int = scheduler_class(**_a ) scheduler.set_timesteps(25 ) _A : List[str] = scheduler.timesteps _A : str = self.dummy_model() _A : Union[str, Any] = self.dummy_sample_deter _A : str = torch.manual_seed(0 ) for i, t in enumerate(_a ): # 1. predict noise residual _A : List[Any] = model(_a , _a ) if i + 1 == timesteps.shape[0]: _A : Dict = None else: _A : Optional[Any] = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 _A : Any = scheduler.step( _a , _a , _a , prev_timestep=_a , generator=_a ).prev_sample _A : int = pred_prev_sample _A : Optional[int] = torch.sum(torch.abs(_a ) ) _A : int = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 258.2044983 ) < 1e-2 assert abs(result_mean.item() - 0.3362038 ) < 1e-3 def a__ ( self ) -> Dict: pass def a__ ( self ) -> Optional[Any]: pass
26
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowercase : int = logging.get_logger(__name__) __lowercase : List[str] = {'vocab_file': 'vocab.json'} __lowercase : List[Any] = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } __lowercase : Tuple = {'mgp-str': 27} class __UpperCamelCase ( lowerCAmelCase_ ): A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __a , __a="[GO]" , __a="[GO]" , __a="[s]" , __a="[GO]" , **__a ): '''simple docstring''' super().__init__( unk_token=__a , bos_token=__a , eos_token=__a , pad_token=__a , **__a , ) with open(__a , encoding='utf-8' ) as vocab_handle: __a : Dict = json.load(__a ) __a : Union[str, Any] = {v: k for k, v in self.vocab.items()} @property def __UpperCAmelCase ( self ): '''simple docstring''' return len(self.vocab ) def __UpperCAmelCase ( self ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __UpperCAmelCase ( self , __a ): '''simple docstring''' __a : str = [] for s in text: char_tokens.extend(__a ) return char_tokens def __UpperCAmelCase ( self , __a ): '''simple docstring''' return self.vocab.get(__a , self.vocab.get(self.unk_token ) ) def __UpperCAmelCase ( self , __a ): '''simple docstring''' return self.decoder.get(__a ) def __UpperCAmelCase ( self , __a , __a = None ): '''simple docstring''' if not os.path.isdir(__a ): logger.error('Vocabulary path ({}) should be a directory'.format(__a ) ) return __a : Any = os.path.join( __a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) with open(__a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=__a , ensure_ascii=__a ) + '\n' ) return (vocab_file,)
27
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
'''simple docstring''' import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _lowerCamelCase : Dict = get_logger(__name__) class SCREAMING_SNAKE_CASE ( enum.Enum ): """simple docstring""" _SCREAMING_SNAKE_CASE = """all_checks""" _SCREAMING_SNAKE_CASE = """basic_checks""" _SCREAMING_SNAKE_CASE = """no_checks""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __lowerCamelCase ( A__ , A__ , A__=None ) -> List[Any]: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(A__ ) - set(A__ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(A__ ) - set(A__ ) ) ) if len(set(A__ ) - set(A__ ) ) > 0: raise UnexpectedDownloadedFile(str(set(A__ ) - set(A__ ) ) ) UpperCamelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] UpperCamelCase = ' for ' + verification_name if verification_name is not None else '' if len(A__ ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __lowerCamelCase ( A__ , A__ ) -> Tuple: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(A__ ) - set(A__ ) ) > 0: raise ExpectedMoreSplits(str(set(A__ ) - set(A__ ) ) ) if len(set(A__ ) - set(A__ ) ) > 0: raise UnexpectedSplits(str(set(A__ ) - set(A__ ) ) ) UpperCamelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(A__ ) > 0: raise NonMatchingSplitsSizesError(str(A__ ) ) logger.info('All the splits matched successfully.' ) def __lowerCamelCase ( A__ , A__ = True ) -> dict: """simple docstring""" if record_checksum: UpperCamelCase = shaaaa() with open(A__ , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'' ): m.update(A__ ) UpperCamelCase = m.hexdigest() else: UpperCamelCase = None return {"num_bytes": os.path.getsize(A__ ), "checksum": checksum} def __lowerCamelCase ( A__ ) -> Optional[int]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
28
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowercase__ ( __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int ): '''simple docstring''' if (ksize % 2) == 0: UpperCAmelCase_ : Optional[Any] = ksize + 1 UpperCAmelCase_ : Optional[int] = np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(__snake_case ): for x in range(__snake_case ): # distance from center UpperCAmelCase_ : Dict = x - ksize // 2 UpperCAmelCase_ : str = y - ksize // 2 # degree to radiant UpperCAmelCase_ : Union[str, Any] = theta / 180 * np.pi UpperCAmelCase_ : Any = np.cos(_theta ) UpperCAmelCase_ : Tuple = np.sin(_theta ) # get kernel x UpperCAmelCase_ : Any = cos_theta * px + sin_theta * py # get kernel y UpperCAmelCase_ : str = -sin_theta * px + cos_theta * py # fill kernel UpperCAmelCase_ : List[str] = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image __UpperCAmelCase = imread('../image_data/lena.jpg') # turn image in gray scale value __UpperCAmelCase = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges __UpperCAmelCase = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 120, 150]: __UpperCAmelCase = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) __UpperCAmelCase = out / out.max() * 255 __UpperCAmelCase = out.astype(np.uinta) imshow('Original', gray) imshow('Gabor filter with 20x20 mask and 6 directions', out) waitKey(0)
29
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class lowercase__: """simple docstring""" def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : Union[str, Any] , ) -> Any: lowercase_ = parent lowercase_ = 1_3 lowercase_ = 7 lowercase_ = True lowercase_ = True lowercase_ = True lowercase_ = 9_9 lowercase_ = 3_2 lowercase_ = 2 lowercase_ = 4 lowercase_ = 3_7 lowercase_ = '''gelu''' lowercase_ = 0.1 lowercase_ = 0.1 lowercase_ = 5_1_2 lowercase_ = 1_6 lowercase_ = 2 lowercase_ = 0.02 lowercase_ = 3 lowercase_ = 4 lowercase_ = None def _lowercase ( self : Optional[int] ) -> Tuple: lowercase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ = None if self.use_input_mask: lowercase_ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ = None lowercase_ = None lowercase_ = None if self.use_labels: lowercase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : Any ) -> Optional[Any]: ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) = self.prepare_config_and_inputs() lowercase_ = True lowercase_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowercase_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _lowercase ( self : Tuple , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : str ) -> List[str]: lowercase_ = TFEsmModel(config=SCREAMING_SNAKE_CASE_ ) lowercase_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} lowercase_ = model(SCREAMING_SNAKE_CASE_ ) lowercase_ = [input_ids, input_mask] lowercase_ = model(SCREAMING_SNAKE_CASE_ ) lowercase_ = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : str , ) -> Union[str, Any]: lowercase_ = True lowercase_ = TFEsmModel(config=SCREAMING_SNAKE_CASE_ ) lowercase_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } lowercase_ = model(SCREAMING_SNAKE_CASE_ ) lowercase_ = [input_ids, input_mask] lowercase_ = model(SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ ) # Also check the case where encoder outputs are not passed lowercase_ = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Any ) -> Any: lowercase_ = TFEsmForMaskedLM(config=SCREAMING_SNAKE_CASE_ ) lowercase_ = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : List[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : str ) -> int: lowercase_ = self.num_labels lowercase_ = TFEsmForTokenClassification(config=SCREAMING_SNAKE_CASE_ ) lowercase_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} lowercase_ = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Tuple ) -> str: lowercase_ = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) = config_and_inputs lowercase_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase__( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): """simple docstring""" a :str = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) a :Union[str, Any] = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) a :Union[str, Any] = False a :Dict = False def _lowercase ( self : List[str] ) -> List[Any]: lowercase_ = TFEsmModelTester(self ) lowercase_ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , hidden_size=3_7 ) def _lowercase ( self : Optional[int] ) -> str: self.config_tester.run_common_tests() def _lowercase ( self : str ) -> int: lowercase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Any ) -> Union[str, Any]: lowercase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : str ) -> Optional[int]: lowercase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Optional[Any] ) -> List[str]: lowercase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*SCREAMING_SNAKE_CASE_ ) @slow def _lowercase ( self : int ) -> List[str]: for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ = TFEsmModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) @unittest.skip('''Protein models do not support embedding resizing.''' ) def _lowercase ( self : str ) -> Tuple: pass @unittest.skip('''Protein models do not support embedding resizing.''' ) def _lowercase ( self : List[Any] ) -> Any: pass def _lowercase ( self : List[str] ) -> Any: lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ = model_class(SCREAMING_SNAKE_CASE_ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer lowercase_ = model.get_bias() assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for k, v in name.items(): assert isinstance(SCREAMING_SNAKE_CASE_ , tf.Variable ) else: lowercase_ = model.get_output_embeddings() assert x is None lowercase_ = model.get_bias() assert name is None @require_tf class lowercase__( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : str ) -> Optional[int]: lowercase_ = TFEsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) lowercase_ = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase_ = model(SCREAMING_SNAKE_CASE_ )[0] lowercase_ = [1, 6, 3_3] self.assertEqual(list(output.numpy().shape ) , SCREAMING_SNAKE_CASE_ ) # compare the actual values for a slice. lowercase_ = tf.constant( [ [ [8.92_15_18, -10.58_98_14, -6.4_67_13_07], [-6.3_96_71_56, -13.91_13_77, -1.1_21_19_15], [-7.78_12_47, -13.95_15_57, -3.74_05_92], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-2 ) ) @slow def _lowercase ( self : int ) -> Union[str, Any]: lowercase_ = TFEsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) lowercase_ = tf.constant([[0, 6, 4, 1_3, 5, 4, 1_6, 1_2, 1_1, 7, 2]] ) lowercase_ = model(SCREAMING_SNAKE_CASE_ )[0] # compare the actual values for a slice. lowercase_ = tf.constant( [ [ [0.14_44_30_92, 0.54_12_53_27, 0.3_24_77_39], [0.30_34_04_84, 0.00_52_66_76, 0.31_07_77_22], [0.32_27_80_43, -0.24_98_70_96, 0.3_41_46_28], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
30
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
'''simple docstring''' from torch import nn def UpperCamelCase_ ( _UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F"""Unsupported activation function: {act_fn}""" )
31
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ : Union[str, Any] = { 'configuration_x_clip': [ 'XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XCLIPConfig', 'XCLIPTextConfig', 'XCLIPVisionConfig', ], 'processing_x_clip': ['XCLIPProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Tuple = [ 'XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'XCLIPModel', 'XCLIPPreTrainedModel', 'XCLIPTextModel', 'XCLIPVisionModel', ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys UpperCAmelCase_ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
32
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
"""simple docstring""" from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = ["vqvae"] def __init__( self : Dict , A : AutoencoderKL , A : UNetaDConditionModel , A : Mel , A : Union[DDIMScheduler, DDPMScheduler] , ) -> List[str]: super().__init__() self.register_modules(unet=A , scheduler=A , mel=A , vqvae=A ) def A ( self : List[str] ) -> int: return 50 if isinstance(self.scheduler , A ) else 10_00 @torch.no_grad() def __call__( self : List[str] , A : int = 1 , A : str = None , A : np.ndarray = None , A : int = 0 , A : int = 0 , A : int = None , A : torch.Generator = None , A : float = 0 , A : float = 0 , A : torch.Generator = None , A : float = 0 , A : torch.Tensor = None , A : torch.Tensor = None , A : Tuple=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: lowercase_ : Dict = steps or self.get_default_steps() self.scheduler.set_timesteps(A ) lowercase_ : List[Any] = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: lowercase_ : List[Any] = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: lowercase_ : int = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=A , device=self.device , ) lowercase_ : Dict = noise lowercase_ : List[Any] = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A , A ) lowercase_ : int = self.mel.audio_slice_to_image(A ) lowercase_ : Tuple = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) lowercase_ : Any = (input_image / 2_55) * 2 - 1 lowercase_ : int = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: lowercase_ : str = self.vqvae.encode(torch.unsqueeze(A , 0 ) ).latent_dist.sample( generator=A )[0] lowercase_ : Optional[Any] = self.vqvae.config.scaling_factor * input_images if start_step > 0: lowercase_ : Any = self.scheduler.add_noise(A , A , self.scheduler.timesteps[start_step - 1] ) lowercase_ : Optional[Any] = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) lowercase_ : Optional[Any] = int(mask_start_secs * pixels_per_second ) lowercase_ : int = int(mask_end_secs * pixels_per_second ) lowercase_ : Union[str, Any] = self.scheduler.add_noise(A , A , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , A ): lowercase_ : Optional[int] = self.unet(A , A , A )['''sample'''] else: lowercase_ : int = self.unet(A , A )['''sample'''] if isinstance(self.scheduler , A ): lowercase_ : List[Any] = self.scheduler.step( model_output=A , timestep=A , sample=A , eta=A , generator=A , )['''prev_sample'''] else: lowercase_ : int = self.scheduler.step( model_output=A , timestep=A , sample=A , generator=A , )['''prev_sample'''] if mask is not None: if mask_start > 0: lowercase_ : List[Any] = mask[:, step, :, :mask_start] if mask_end > 0: lowercase_ : List[str] = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance lowercase_ : Optional[Any] = 1 / self.vqvae.config.scaling_factor * images lowercase_ : str = self.vqvae.decode(A )['''sample'''] lowercase_ : Tuple = (images / 2 + 0.5).clamp(0 , 1 ) lowercase_ : Optional[Any] = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() lowercase_ : List[str] = (images * 2_55).round().astype('''uint8''' ) lowercase_ : Optional[int] = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A , mode='''RGB''' ).convert('''L''' ) for _ in images) ) lowercase_ : Any = [self.mel.image_to_audio(A ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A )[:, np.newaxis, :] ) , **ImagePipelineOutput(A ) ) @torch.no_grad() def A ( self : List[str] , A : List[Image.Image] , A : int = 50 ) -> np.ndarray: assert isinstance(self.scheduler , A ) self.scheduler.set_timesteps(A ) lowercase_ : List[str] = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) lowercase_ : Union[str, Any] = (sample / 2_55) * 2 - 1 lowercase_ : Union[str, Any] = torch.Tensor(A ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): lowercase_ : Union[str, Any] = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps lowercase_ : Dict = self.scheduler.alphas_cumprod[t] lowercase_ : int = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) lowercase_ : str = 1 - alpha_prod_t lowercase_ : Dict = self.unet(A , A )['''sample'''] lowercase_ : List[Any] = (1 - alpha_prod_t_prev) ** 0.5 * model_output lowercase_ : List[str] = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) lowercase_ : int = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def A ( A : torch.Tensor , A : torch.Tensor , A : float ) -> torch.Tensor: lowercase_ : List[str] = acos(torch.dot(torch.flatten(A ) , torch.flatten(A ) ) / torch.norm(A ) / torch.norm(A ) ) return sin((1 - alpha) * theta ) * xa / sin(A ) + sin(alpha * theta ) * xa / sin(A )
33
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0
'''simple docstring''' import math def snake_case_ (_a : float , _a : float ): return math.pow(_a , 2 ) - a def snake_case_ (_a : float ): return 2 * x def snake_case_ (_a : float ): UpperCAmelCase = 2.0 while start <= a: UpperCAmelCase = math.pow(_a , 2 ) return start def snake_case_ (_a : float , _a : int = 9_9_9_9 , _a : float = 0.00_0000_0000_0001 ): if a < 0: raise ValueError('''math domain error''' ) UpperCAmelCase = get_initial_point(_a ) for _ in range(_a ): UpperCAmelCase = value UpperCAmelCase = value - fx(_a , _a ) / fx_derivative(_a ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
34
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
'''simple docstring''' import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py __a = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evaluation of Machine Translation},\n booktitle = {},\n year = {2002},\n pages = {311--318}\n}\n@inproceedings{lin-och-2004-orange,\n title = \"{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation\",\n author = \"Lin, Chin-Yew and\n Och, Franz Josef\",\n booktitle = \"{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics\",\n month = \"aug 23{--}aug 27\",\n year = \"2004\",\n address = \"Geneva, Switzerland\",\n publisher = \"COLING\",\n url = \"https://www.aclweb.org/anthology/C04-1072\",\n pages = \"501--507\",\n}\n" __a = "\\nBLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.\nQuality is considered to be the correspondence between a machine's output and that of a human: \"the closer a machine translation is to a professional human translation,\nthe better it is\" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and\nremains one of the most popular automated and inexpensive metrics.\n\nScores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.\nThose scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality. Intelligibility or grammatical correctness\nare not taken into account[citation needed].\n\nBLEU's output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1\nrepresenting more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the\nreference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional\nreference translations will increase the BLEU score.\n" __a = "\nComputes BLEU score of translated segments against one or more references.\nArgs:\n predictions: list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n max_order: Maximum n-gram order to use when computing BLEU score.\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\nReturns:\n 'bleu': bleu score,\n 'precisions': geometric mean of n-gram precisions,\n 'brevity_penalty': brevity penalty,\n 'length_ratio': ratio of lengths,\n 'translation_length': translation_length,\n 'reference_length': reference_length\nExamples:\n\n >>> predictions = [\n ... [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample\n ... [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample\n ... ]\n >>> references = [\n ... [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references)\n ... [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference)\n ... ]\n >>> bleu = datasets.load_metric(\"bleu\")\n >>> results = bleu.compute(predictions=predictions, references=references)\n >>> print(results[\"bleu\"])\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): """simple docstring""" def lowerCamelCase ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ), """references""": datasets.Sequence( datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py"""] , reference_urls=[ """https://en.wikipedia.org/wiki/BLEU""", """https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213""", ] , ) def lowerCamelCase ( self : Optional[int] , snake_case_ : List[Any] , snake_case_ : int , snake_case_ : int=4 , snake_case_ : List[Any]=False ): snake_case__ : Optional[Any] = compute_bleu( reference_corpus=snake_case_ , translation_corpus=snake_case_ , max_order=snake_case_ , smooth=snake_case_ ) ((snake_case__) , (snake_case__) , (snake_case__) , (snake_case__) , (snake_case__) , (snake_case__)) : Optional[int] = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
35
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
import re def A ( _lowerCamelCase ): '''simple docstring''' return [char.split() for char in re.split(r"[^ a-z A-Z 0-9 \s]" , str_ )] def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Tuple = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' try: _lowerCAmelCase : Any = split_input(_lowerCamelCase ) if upper: _lowerCAmelCase : Dict = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: _lowerCAmelCase : Dict = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def A ( _lowerCamelCase ): '''simple docstring''' return to_simple_case(_lowerCamelCase ) def A ( _lowerCamelCase ): '''simple docstring''' try: _lowerCAmelCase : Dict = to_simple_case(_lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' return to_complex_case(_lowerCamelCase , _lowerCamelCase , "_" ) def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' return to_complex_case(_lowerCamelCase , _lowerCamelCase , "-" ) if __name__ == "__main__": __import__("doctest").testmod()
36
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowercase : Optional[Any] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") lowercase : Tuple = parser.parse_args() lowercase : Optional[int] = "cpu" lowercase : Optional[Any] = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowercase : Optional[int] = "path-to-your-trained-model" lowercase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowercase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowercase : Dict = pipe.to(device) # to channels last lowercase : Optional[Any] = pipe.unet.to(memory_format=torch.channels_last) lowercase : int = pipe.vae.to(memory_format=torch.channels_last) lowercase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowercase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowercase : Any = torch.randn(2, 4, 64, 64) lowercase : Optional[int] = torch.rand(1) * 999 lowercase : Optional[Any] = torch.randn(2, 77, 768) lowercase : Optional[Any] = (sample, timestep, encoder_hidden_status) try: lowercase : List[Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowercase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Optional[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowercase : Tuple = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowercase : List[str] = 666 lowercase : Tuple = torch.Generator(device).manual_seed(seed) lowercase : Union[str, Any] = {"generator": generator} if args.steps is not None: lowercase : Dict = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowercase : List[str] = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
42
0
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' @staticmethod @abstractmethod def UpperCAmelCase_ ( __UpperCAmelCase ) -> Optional[Any]: raise NotImplementedError() @abstractmethod def UpperCAmelCase_ ( self ) -> int: raise NotImplementedError()
37
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCAmelCase ( _lowerCamelCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return 0.0 def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[int | float, int | float]: _snake_case = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.abs(np.fft.fft(__A ) ) _snake_case = 20 * np.logaa(__A ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds _snake_case = get_bounds(__A , __A ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__A ) plt.show() def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.angle(np.fft.fft(__A ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__A , -2 * pi ) ) plt.show()
42
0
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Dict ) -> Any: """simple docstring""" return 1 / (1 + np.exp(-z )) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[Any] , __magic_name__ : Dict ) -> str: """simple docstring""" return (-y * np.log(__magic_name__ ) - (1 - y) * np.log(1 - h )).mean() def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Tuple , __magic_name__ : List[Any] , __magic_name__ : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase :Tuple = np.dot(__magic_name__ , __magic_name__ ) return np.sum(y * scores - np.log(1 + np.exp(__magic_name__ ) ) ) def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str , __magic_name__ : List[str] , __magic_name__ : Optional[int] , __magic_name__ : Optional[Any]=7_0000 ) -> Optional[Any]: """simple docstring""" UpperCamelCase :Tuple = np.zeros(x.shape[1] ) for iterations in range(__magic_name__ ): UpperCamelCase :str = np.dot(__magic_name__ , __magic_name__ ) UpperCamelCase :str = sigmoid_function(__magic_name__ ) UpperCamelCase :int = np.dot(x.T , h - y ) / y.size UpperCamelCase :int = theta - alpha * gradient # updating the weights UpperCamelCase :List[str] = np.dot(__magic_name__ , __magic_name__ ) UpperCamelCase :str = sigmoid_function(__magic_name__ ) UpperCamelCase :List[Any] = cost_function(__magic_name__ , __magic_name__ ) if iterations % 100 == 0: print(f"""loss: {j} \t""" ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": UpperCAmelCase_ : int = datasets.load_iris() UpperCAmelCase_ : Dict = iris.data[:, :2] UpperCAmelCase_ : Dict = (iris.target != 0) * 1 UpperCAmelCase_ : Optional[int] = 0.1 UpperCAmelCase_ : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_00_00) print('''theta: ''', theta) # printing the theta i.e our weights vector def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[int] ) -> List[str]: """simple docstring""" return sigmoid_function( np.dot(__magic_name__ , __magic_name__ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((UpperCAmelCase_) , (UpperCAmelCase_)) : Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((UpperCAmelCase_) , (UpperCAmelCase_)) : List[str] = (x[:, 1].min(), x[:, 1].max()) ((UpperCAmelCase_) , (UpperCAmelCase_)) : List[str] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) UpperCAmelCase_ : List[Any] = np.c_[xxa.ravel(), xxa.ravel()] UpperCAmelCase_ : Tuple = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
38
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class __UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = vocab_size _snake_case = d_embed _snake_case = d_proj _snake_case = cutoffs + [vocab_size] _snake_case = [0] + self.cutoffs _snake_case = div_val _snake_case = self.cutoffs[0] _snake_case = len(self.cutoffs ) - 1 _snake_case = self.shortlist_size + self.n_clusters _snake_case = keep_order _snake_case = [] _snake_case = [] def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if self.n_clusters > 0: _snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_weight' ) _snake_case = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: _snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' , ) self.out_projs.append(lowerCAmelCase_ ) else: self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case = self.d_embed // (self.div_val**i) _snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' ) self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = x if proj is not None: _snake_case = tf.einsum('ibd,ed->ibe' , lowerCAmelCase_ , lowerCAmelCase_ ) return tf.einsum('ibd,nd->ibn' , lowerCAmelCase_ , lowerCAmelCase_ ) + b @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = tf.range(lp_size[0] , dtype=target.dtype ) _snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False ): """simple docstring""" _snake_case = 0 if self.n_clusters == 0: _snake_case = self._logit(lowerCAmelCase_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: _snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase_ , logits=lowerCAmelCase_ ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ , axis=-1 ) else: _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = [] _snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: _snake_case = (target >= l_idx) & (target < r_idx) _snake_case = tf.where(lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) - l_idx if self.div_val == 1: _snake_case = self.out_layers[0][0][l_idx:r_idx] _snake_case = self.out_layers[0][1][l_idx:r_idx] else: _snake_case = self.out_layers[i][0] _snake_case = self.out_layers[i][1] if i == 0: _snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) _snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[0] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) else: _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[i] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) _snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster _snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase_ ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase_ , -cur_logprob , shape_list(lowerCAmelCase_ ) ) _snake_case = tf.concat(lowerCAmelCase_ , axis=-1 ) if target is not None: if return_mean: _snake_case = tf.reduce_mean(lowerCAmelCase_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase_ , name=self.name , aggregation='mean' if return_mean else '' ) return out
42
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=7 , UpperCAmelCase=3 , UpperCAmelCase=18 , UpperCAmelCase=30 , UpperCAmelCase=400 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=[0.5, 0.5, 0.5] , UpperCAmelCase=[0.5, 0.5, 0.5] , ): """simple docstring""" _UpperCAmelCase = size if size is not None else {'shortest_edge': 18} _UpperCAmelCase = crop_size if crop_size is not None else {'height': 18, 'width': 18} _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = num_channels _UpperCAmelCase = image_size _UpperCAmelCase = min_resolution _UpperCAmelCase = max_resolution _UpperCAmelCase = do_resize _UpperCAmelCase = size _UpperCAmelCase = do_center_crop _UpperCAmelCase = crop_size _UpperCAmelCase = do_normalize _UpperCAmelCase = image_mean _UpperCAmelCase = image_std def UpperCamelCase ( self ): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = LevitImageProcessor if is_vision_available() else None def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = LevitImageProcessingTester(self ) @property def UpperCamelCase ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'image_std' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_center_crop' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'size' ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) _UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , Image.Image ) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _UpperCAmelCase = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , numpify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , np.ndarray ) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _UpperCAmelCase = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , torchify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , torch.Tensor ) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _UpperCAmelCase = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
39
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
0
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class _A ( unittest.TestCase ): """simple docstring""" def __snake_case ( self : Optional[Any] , __UpperCAmelCase : str): a : Optional[int] = 3 a : List[Any] = 250 a : Optional[Any] = ids_tensor((batch_size, length) , __UpperCAmelCase) a : List[Any] = torch.ones((batch_size, length) , device=__UpperCAmelCase , dtype=torch.float) / length return input_ids, scores def __snake_case ( self : Union[str, Any]): a , a : List[Any] = self._get_tensors(5) a : Union[str, Any] = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10), MaxTimeCriteria(max_time=0.1), ]) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : Any = self._get_tensors(9) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : Any = self._get_tensors(10) self.assertTrue(criteria(__UpperCAmelCase , __UpperCAmelCase)) def __snake_case ( self : Union[str, Any]): a : Any = MaxLengthCriteria(max_length=10) a , a : Optional[Any] = self._get_tensors(5) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : List[Any] = self._get_tensors(9) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : List[Any] = self._get_tensors(10) self.assertTrue(criteria(__UpperCAmelCase , __UpperCAmelCase)) def __snake_case ( self : Optional[Any]): a : str = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5) a , a : str = self._get_tensors(5) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : str = self._get_tensors(9) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a , a : Dict = self._get_tensors(10) self.assertTrue(criteria(__UpperCAmelCase , __UpperCAmelCase)) a : Dict = StoppingCriteriaList([criteria]) self.assertEqual(criteria_list.max_length , 10) def __snake_case ( self : str): a , a : Dict = self._get_tensors(5) a : List[str] = MaxTimeCriteria(max_time=0.1) self.assertFalse(criteria(__UpperCAmelCase , __UpperCAmelCase)) a : Dict = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2) self.assertTrue(criteria(__UpperCAmelCase , __UpperCAmelCase)) def __snake_case ( self : Union[str, Any]): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10)]) , 10) with self.assertWarns(__UpperCAmelCase): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10)]) , 11) a : Tuple = validate_stopping_criteria(StoppingCriteriaList() , 11) self.assertEqual(len(__UpperCAmelCase) , 1)
40
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot import BlenderbotTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _A : List[Any] =logging.get_logger(__name__) _A : Tuple ={ '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _A : List[Any] ={ '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } _A : List[Any] ={'''facebook/blenderbot-3B''': 128} class _lowercase ( _lowercase ): a = VOCAB_FILES_NAMES a = PRETRAINED_VOCAB_FILES_MAP a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a = ["""input_ids""", """attention_mask"""] a = BlenderbotTokenizer def __init__( self: Union[str, Any] , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[str]=None , UpperCamelCase__: int=None , UpperCamelCase__: Dict="replace" , UpperCamelCase__: Any="<s>" , UpperCamelCase__: Dict="</s>" , UpperCamelCase__: Any="</s>" , UpperCamelCase__: Union[str, Any]="<s>" , UpperCamelCase__: Tuple="<unk>" , UpperCamelCase__: Union[str, Any]="<pad>" , UpperCamelCase__: Optional[Any]="<mask>" , UpperCamelCase__: Tuple=False , UpperCamelCase__: str=True , **UpperCamelCase__: Optional[Any] , ): super().__init__( UpperCamelCase__ , UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , errors=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase__ : Optional[int] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCamelCase__ ) != add_prefix_space: lowerCamelCase__ : List[str] = getattr(UpperCamelCase__ , pre_tok_state.pop("""type""" ) ) lowerCamelCase__ : Union[str, Any] = add_prefix_space lowerCamelCase__ : Optional[int] = pre_tok_class(**UpperCamelCase__ ) lowerCamelCase__ : Dict = add_prefix_space lowerCamelCase__ : Tuple = """post_processor""" lowerCamelCase__ : Tuple = getattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__ ) if tokenizer_component_instance: lowerCamelCase__ : str = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : str = tuple(state["""sep"""] ) if "cls" in state: lowerCamelCase__ : Optional[Any] = tuple(state["""cls"""] ) lowerCamelCase__ : Optional[int] = False if state.get("""add_prefix_space""" , UpperCamelCase__ ) != add_prefix_space: lowerCamelCase__ : Tuple = add_prefix_space lowerCamelCase__ : Optional[Any] = True if state.get("""trim_offsets""" , UpperCamelCase__ ) != trim_offsets: lowerCamelCase__ : int = trim_offsets lowerCamelCase__ : int = True if changes_to_apply: lowerCamelCase__ : List[Any] = getattr(UpperCamelCase__ , state.pop("""type""" ) ) lowerCamelCase__ : Any = component_class(**UpperCamelCase__ ) setattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__ ) @property # Copied from transformers.models.roberta.tokenization_roberta_fast.RobertaTokenizerFast.mask_token with Roberta->Blenderbot, RoBERTa->Blenderbot def lowerCamelCase_ ( self: str ): if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""" ) return None return str(self._mask_token ) @mask_token.setter def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else value lowerCamelCase__ : int = value def lowerCamelCase_ ( self: Union[str, Any] , *UpperCamelCase__: Optional[Any] , **UpperCamelCase__: Any ): lowerCamelCase__ : List[str] = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] , *UpperCamelCase__: List[Any] , **UpperCamelCase__: int ): lowerCamelCase__ : List[str] = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Optional[str] = None ): lowerCamelCase__ : Dict = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): return token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self: Any , UpperCamelCase__: "Conversation" ): lowerCamelCase__ : List[str] = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text ) else: # Generated responses should contain them already. inputs.append(UpperCamelCase__ ) lowerCamelCase__ : str = """ """.join(UpperCamelCase__ ) lowerCamelCase__ : str = self.encode(UpperCamelCase__ ) if len(UpperCamelCase__ ) > self.model_max_length: lowerCamelCase__ : List[Any] = input_ids[-self.model_max_length :] logger.warning(F'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''' ) return input_ids
41
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = 1 _snake_case = 2 while i * i <= n: _snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _snake_case = 1 _snake_case = 1 while True: i += 1 t_num += i if count_divisors(__A ) > 500: break return t_num if __name__ == "__main__": print(solution())
42
0
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __UpperCAmelCase ( _lowerCamelCase ): @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _snake_case = self.get_env() _snake_case = '1' _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import AutoModel\n ' _snake_case = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
42
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __A : def __init__( self , a__ , a__=13 , a__=7 , a__=True , a__=True , a__=True , a__=True , a__=99 , a__=[1, 1, 2] , a__=1 , a__=32 , a__=4 , a__=8 , a__=37 , a__="gelu_new" , a__=0.1 , a__=0.1 , a__=0.0 , a__=512 , a__=3 , a__=0.0_2 , a__=3 , a__=4 , a__=None , a__=False , ): _lowerCAmelCase : Union[str, Any] = parent _lowerCAmelCase : Optional[Any] = batch_size _lowerCAmelCase : Optional[Any] = seq_length _lowerCAmelCase : Union[str, Any] = is_training _lowerCAmelCase : List[Any] = use_input_mask _lowerCAmelCase : List[Any] = use_token_type_ids _lowerCAmelCase : List[Any] = use_labels _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Any = block_sizes _lowerCAmelCase : List[str] = num_decoder_layers _lowerCAmelCase : List[Any] = d_model _lowerCAmelCase : List[Any] = n_head _lowerCAmelCase : Any = d_head _lowerCAmelCase : Union[str, Any] = d_inner _lowerCAmelCase : Optional[int] = hidden_act _lowerCAmelCase : Optional[Any] = hidden_dropout _lowerCAmelCase : Optional[int] = attention_dropout _lowerCAmelCase : int = activation_dropout _lowerCAmelCase : str = max_position_embeddings _lowerCAmelCase : Optional[Any] = type_vocab_size _lowerCAmelCase : Optional[Any] = 2 _lowerCAmelCase : List[str] = num_labels _lowerCAmelCase : Dict = num_choices _lowerCAmelCase : Optional[int] = scope _lowerCAmelCase : List[str] = initializer_std # Used in the tests to check the size of the first attention layer _lowerCAmelCase : Dict = n_head # Used in the tests to check the size of the first hidden state _lowerCAmelCase : Union[str, Any] = self.d_model # Used in the tests to check the number of output hidden states/attentions _lowerCAmelCase : List[Any] = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: _lowerCAmelCase : List[Any] = self.num_hidden_layers + 2 def __A ( self ): _lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : Dict = None if self.use_input_mask: _lowerCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase : Dict = None if self.use_token_type_ids: _lowerCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : Dict = None _lowerCAmelCase : Dict = None _lowerCAmelCase : Tuple = None if self.use_labels: _lowerCAmelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCAmelCase : Tuple = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : List[str] = TFFunnelModel(config=a__ ) _lowerCAmelCase : Any = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : List[Any] = model(a__ ) _lowerCAmelCase : Tuple = [input_ids, input_mask] _lowerCAmelCase : List[str] = model(a__ ) _lowerCAmelCase : Any = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) _lowerCAmelCase : str = False _lowerCAmelCase : List[str] = TFFunnelModel(config=a__ ) _lowerCAmelCase : Optional[int] = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) _lowerCAmelCase : int = False _lowerCAmelCase : Tuple = TFFunnelModel(config=a__ ) _lowerCAmelCase : int = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Optional[Any] = TFFunnelBaseModel(config=a__ ) _lowerCAmelCase : Dict = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : Any = model(a__ ) _lowerCAmelCase : Optional[int] = [input_ids, input_mask] _lowerCAmelCase : Any = model(a__ ) _lowerCAmelCase : Optional[int] = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) _lowerCAmelCase : List[Any] = False _lowerCAmelCase : int = TFFunnelBaseModel(config=a__ ) _lowerCAmelCase : List[str] = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) _lowerCAmelCase : Any = False _lowerCAmelCase : Optional[Any] = TFFunnelBaseModel(config=a__ ) _lowerCAmelCase : Dict = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Union[str, Any] = TFFunnelForPreTraining(config=a__ ) _lowerCAmelCase : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : Optional[int] = model(a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Any = TFFunnelForMaskedLM(config=a__ ) _lowerCAmelCase : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : List[str] = model(a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[str] = TFFunnelForSequenceClassification(config=a__ ) _lowerCAmelCase : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : List[Any] = model(a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : List[str] = self.num_choices _lowerCAmelCase : Dict = TFFunnelForMultipleChoice(config=a__ ) _lowerCAmelCase : Any = tf.tile(tf.expand_dims(a__ , 1 ) , (1, self.num_choices, 1) ) _lowerCAmelCase : int = tf.tile(tf.expand_dims(a__ , 1 ) , (1, self.num_choices, 1) ) _lowerCAmelCase : str = tf.tile(tf.expand_dims(a__ , 1 ) , (1, self.num_choices, 1) ) _lowerCAmelCase : List[str] = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } _lowerCAmelCase : List[str] = model(a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Tuple = self.num_labels _lowerCAmelCase : List[Any] = TFFunnelForTokenClassification(config=a__ ) _lowerCAmelCase : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : Dict = model(a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __A ( self , a__ , a__ , a__ , a__ , a__ , a__ , a__ , ): _lowerCAmelCase : Tuple = TFFunnelForQuestionAnswering(config=a__ ) _lowerCAmelCase : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _lowerCAmelCase : Optional[Any] = model(a__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __A ( self ): _lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCAmelCase : str = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __A ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Dict = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) _UpperCamelCase : int = ( { "feature-extraction": (TFFunnelBaseModel, TFFunnelModel), "fill-mask": TFFunnelForMaskedLM, "question-answering": TFFunnelForQuestionAnswering, "text-classification": TFFunnelForSequenceClassification, "token-classification": TFFunnelForTokenClassification, "zero-shot": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) _UpperCamelCase : List[str] = False _UpperCamelCase : List[str] = False def __A ( self ): _lowerCAmelCase : Dict = TFFunnelModelTester(self ) _lowerCAmelCase : Union[str, Any] = ConfigTester(self , config_class=a__ ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def __A ( self ): _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*a__ ) def __A ( self ): _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*a__ ) def __A ( self ): _lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a__ ) def __A ( self ): _lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a__ ) @require_tf class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : List[str] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) _UpperCamelCase : Optional[Any] = False _UpperCamelCase : Optional[Any] = False def __A ( self ): _lowerCAmelCase : Tuple = TFFunnelModelTester(self , base=a__ ) _lowerCAmelCase : List[Any] = ConfigTester(self , config_class=a__ ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*a__ ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*a__ )
44
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCAmelCase_ ).to(lowerCAmelCase_ ) _snake_case = AutoTokenizer.from_pretrained('google/mt5-small' ) _snake_case = tokenizer('Hello there' , return_tensors='pt' ).input_ids _snake_case = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _snake_case = model(input_ids.to(lowerCAmelCase_ ) , labels=labels.to(lowerCAmelCase_ ) ).loss _snake_case = -(labels.shape[-1] * loss.item()) _snake_case = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
42
0
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase_ = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def lowercase ( lowerCAmelCase__ : Optional[Any]=None ) -> Optional[int]: if subparsers is not None: __a = subparsers.add_parser('''tpu-config''' , description=_description ) else: __a = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments __a = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) __a = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def lowercase ( lowerCAmelCase__ : Optional[Any] ) -> Optional[Any]: __a = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): __a = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: __a = defaults.command_file if not args.command and defaults.commands is not None: __a = defaults.commands if not args.tpu_name: __a = defaults.tpu_name if not args.tpu_zone: __a = defaults.tpu_zone if args.accelerate_version == "dev": __a = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": __a = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): __a = f'''accelerate=={args.accelerate_version}''' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: __a = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): __a = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate __a = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'''pip install {args.accelerate_version}'''] new_cmd += args.command __a = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess __a = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'''Running {' '.join(lowerCAmelCase__ )}''' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def lowercase ( ) -> str: __a = tpu_command_parser() __a = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
45
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[str] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE__ = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece class lowercase ( _UpperCAmelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE = XLMProphetNetTokenizer _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = True def _snake_case ( self ) -> str: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase = XLMProphetNetTokenizer(lowercase , keep_accents=lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self ) -> Tuple: lowerCAmelCase = """[PAD]""" lowerCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase ) , lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase ) , lowercase ) def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(lowercase ) , 1_012 ) def _snake_case ( self ) -> int: self.assertEqual(self.get_tokenizer().vocab_size , 1_012 ) def _snake_case ( self ) -> Tuple: lowerCAmelCase = XLMProphetNetTokenizer(lowercase , keep_accents=lowercase ) lowerCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(lowercase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) lowerCAmelCase = tokenizer.convert_tokens_to_ids(lowercase ) self.assertListEqual( lowercase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) lowerCAmelCase = tokenizer.convert_ids_to_tokens(lowercase ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def _snake_case ( self ) -> Dict: return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def _snake_case ( self ) -> int: lowerCAmelCase = """Hello World!""" lowerCAmelCase = [35_389, 6_672, 49, 2] self.assertListEqual(lowercase , self.big_tokenizer.encode(lowercase ) ) @slow def _snake_case ( self ) -> Any: # fmt: off lowerCAmelCase = {"""input_ids""": [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
46
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowercase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowercase : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowercase : set[int] = {ord(char) for char in VALID_CHARS} lowercase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str | None: _snake_case = "" _snake_case = 42 _snake_case = 42 _snake_case = 42 for keychar, cipherchar in zip(cycle(__A ) , __A ): _snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__A ) return decoded def SCREAMING_SNAKE_CASE__ ( __A ) -> list[str]: _snake_case = [] for key in product(__A , repeat=3 ): _snake_case = try_key(__A , __A ) if encoded is not None: possibles.append(__A ) return possibles def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __A = "p059_cipher.txt" ) -> int: _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = Path(__A ).parent.joinpath(__A ).read_text(encoding='utf-8' ) _snake_case = [int(__A ) for number in data.strip().split(',' )] _snake_case = filter_valid_chars(__A ) for common_word in COMMON_WORDS: _snake_case = filter_common_word(__A , __A ) if len(__A ) == 1: break _snake_case = possibles[0] return sum(ord(__A ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' def _lowerCAmelCase ( _UpperCamelCase : str , _UpperCamelCase : str ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE =len(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =len(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =[[False for _ in range(m + 1 )] for _ in range(n + 1 )] _SCREAMING_SNAKE_CASE =True for i in range(_UpperCamelCase ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _SCREAMING_SNAKE_CASE =True if a[i].islower(): _SCREAMING_SNAKE_CASE =True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
47
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
import string def A ( _SCREAMING_SNAKE_CASE ) -> None: for key in range(len(string.ascii_uppercase ) ): lowerCamelCase : Optional[int] = "" for symbol in message: if symbol in string.ascii_uppercase: lowerCamelCase : Any = string.ascii_uppercase.find(_SCREAMING_SNAKE_CASE ) lowerCamelCase : Optional[int] = num - key if num < 0: lowerCamelCase : Union[str, Any] = num + len(string.ascii_uppercase ) lowerCamelCase : str = translated + string.ascii_uppercase[num] else: lowerCamelCase : Optional[Any] = translated + symbol print(f'''Decryption using Key #{key}: {translated}''' ) def A ( ) -> None: lowerCamelCase : List[Any] = input("Encrypted message: " ) lowerCamelCase : int = message.upper() decrypt(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() main()
48
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case :List[Any] = logging.get_logger(__name__) __snake_case :Dict = { '''google/mobilenet_v2_1.4_224''': '''https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json''', '''google/mobilenet_v2_1.0_224''': '''https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json''', '''google/mobilenet_v2_0.75_160''': '''https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json''', '''google/mobilenet_v2_0.35_96''': '''https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json''', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class _A ( __UpperCAmelCase ): UpperCamelCase__ : Tuple = '''mobilenet_v2''' def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Tuple=3 , __SCREAMING_SNAKE_CASE : int=224 , __SCREAMING_SNAKE_CASE : Optional[int]=1.0 , __SCREAMING_SNAKE_CASE : Union[str, Any]=8 , __SCREAMING_SNAKE_CASE : List[Any]=8 , __SCREAMING_SNAKE_CASE : int=6 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : List[Any]=True , __SCREAMING_SNAKE_CASE : Dict=True , __SCREAMING_SNAKE_CASE : Dict="relu6" , __SCREAMING_SNAKE_CASE : Dict=True , __SCREAMING_SNAKE_CASE : Dict=0.8 , __SCREAMING_SNAKE_CASE : Dict=0.02 , __SCREAMING_SNAKE_CASE : Dict=0.0_01 , __SCREAMING_SNAKE_CASE : Dict=255 , **__SCREAMING_SNAKE_CASE : Optional[int] , ): '''simple docstring''' super().__init__(**__SCREAMING_SNAKE_CASE) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''') __a = num_channels __a = image_size __a = depth_multiplier __a = depth_divisible_by __a = min_depth __a = expand_ratio __a = output_stride __a = first_layer_is_expansion __a = finegrained_output __a = hidden_act __a = tf_padding __a = classifier_dropout_prob __a = initializer_range __a = layer_norm_eps __a = semantic_loss_ignore_index class _A ( __UpperCAmelCase ): UpperCamelCase__ : Any = version.parse('''1.11''' ) @property def _lowerCamelCase ( self : Dict): '''simple docstring''' return OrderedDict([('''pixel_values''', {0: '''batch'''})]) @property def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})]) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})]) @property def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' return 1E-4
49
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
def A (__A : int , __A : bool = False ) -> bool: """simple docstring""" if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( '''Warning: upper bound of deterministic test is exceeded. ''' '''Pass allow_probable=True to allow probabilistic test. ''' '''A return value of True indicates a probable prime.''' ) # array bounds provided by analysis UpperCAmelCase_ = [ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] UpperCAmelCase_ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(__A , 1 ): if n < _p: # then we have our last prime to check UpperCAmelCase_ = primes[:idx] break UpperCAmelCase_ , UpperCAmelCase_ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: UpperCAmelCase_ = False for r in range(__A ): UpperCAmelCase_ = pow(__A , d * 2**r , __A ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): UpperCAmelCase_ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def A () -> None: """simple docstring""" assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
51
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
import qiskit def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> qiskit.result.counts.Counts: UpperCamelCase : List[str] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register UpperCamelCase : List[Any] = qiskit.QuantumCircuit(_lowerCAmelCase , _lowerCAmelCase ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] , [0, 1] ) # Execute the circuit on the qasm simulator UpperCamelCase : Optional[int] = qiskit.execute(_lowerCAmelCase , _lowerCAmelCase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Optional[Any] = single_qubit_measure(2, 2) print(f"""Total count for various states are: {counts}""")
52
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
'''simple docstring''' a__ : List[Any] ={0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} a__ : Optional[int] ={0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowercase__ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: """simple docstring""" __UpperCamelCase = True __UpperCamelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__lowercase , __lowercase , __lowercase ) order.append(__lowercase ) return order def lowercase__ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: """simple docstring""" __UpperCamelCase = True __UpperCamelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__lowercase , __lowercase , __lowercase ) return component def lowercase__ ( __lowercase : dict[int, list[int]] ) -> list[list[int]]: """simple docstring""" __UpperCamelCase = len(__lowercase ) * [False] __UpperCamelCase = {vert: [] for vert in range(len(__lowercase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__lowercase ) __UpperCamelCase = [] for i, was_visited in enumerate(__lowercase ): if not was_visited: order += topology_sort(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = [] __UpperCamelCase = len(__lowercase ) * [False] for i in range(len(__lowercase ) ): __UpperCamelCase = order[len(__lowercase ) - i - 1] if not visited[vert]: __UpperCamelCase = find_components(__lowercase , __lowercase , __lowercase ) components_list.append(__lowercase ) return components_list
53
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Optional[int] = "philschmid/bart-large-cnn-samsum" snake_case__ : str = ( "This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, " "and returns a summary of the text." ) snake_case__ : Union[str, Any] = "summarizer" snake_case__ : Optional[Any] = AutoTokenizer snake_case__ : Optional[int] = AutoModelForSeqaSeqLM snake_case__ : Any = ["text"] snake_case__ : Tuple = ["text"] def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : Optional[Any] ) -> int: return self.pre_processor(UpperCAmelCase__ , return_tensors="pt" , truncation=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : int ) -> Any: return self.model.generate(**UpperCAmelCase__ )[0] def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : List[Any] ) -> List[Any]: return self.pre_processor.decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ , clean_up_tokenization_spaces=UpperCAmelCase__ )
54
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class snake_case ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained("hf-internal-testing/tiny-random-GPT2Model" ) lowerCamelCase_ = BertTokenizerFast.from_pretrained("hf-internal-testing/tiny-random-bert" ) lowerCamelCase_ = InstructBlipProcessor(UpperCamelCase , UpperCamelCase , UpperCamelCase ) processor.save_pretrained(self.tmpdirname ) def snake_case ( self , **UpperCamelCase ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase ).tokenizer def snake_case ( self , **UpperCamelCase ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase ).image_processor def snake_case ( self , **UpperCamelCase ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase ).qformer_tokenizer def snake_case ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(UpperCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def snake_case ( self ): """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) lowerCamelCase_ = self.get_image_processor(do_normalize=UpperCamelCase , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=UpperCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase ) self.assertIsInstance(processor.qformer_tokenizer , UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=UpperCamelCase , image_processor=UpperCamelCase , qformer_tokenizer=UpperCamelCase ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(UpperCamelCase , return_tensors="np" ) lowerCamelCase_ = processor(images=UpperCamelCase , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=UpperCamelCase , image_processor=UpperCamelCase , qformer_tokenizer=UpperCamelCase ) lowerCamelCase_ = "lower newer" lowerCamelCase_ = processor(text=UpperCamelCase ) lowerCamelCase_ = tokenizer(UpperCamelCase , return_token_type_ids=UpperCamelCase ) lowerCamelCase_ = qformer_tokenizer(UpperCamelCase , return_token_type_ids=UpperCamelCase ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor["qformer_" + key] ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=UpperCamelCase , image_processor=UpperCamelCase , qformer_tokenizer=UpperCamelCase ) lowerCamelCase_ = "lower newer" lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=UpperCamelCase , images=UpperCamelCase ) self.assertListEqual( list(inputs.keys() ) , ["input_ids", "attention_mask", "qformer_input_ids", "qformer_attention_mask", "pixel_values"] , ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase ): processor() def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=UpperCamelCase , image_processor=UpperCamelCase , qformer_tokenizer=UpperCamelCase ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(UpperCamelCase ) lowerCamelCase_ = tokenizer.batch_decode(UpperCamelCase ) self.assertListEqual(UpperCamelCase , UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=UpperCamelCase , image_processor=UpperCamelCase , qformer_tokenizer=UpperCamelCase ) lowerCamelCase_ = "lower newer" lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=UpperCamelCase , images=UpperCamelCase ) self.assertListEqual( list(inputs.keys() ) , ["input_ids", "attention_mask", "qformer_input_ids", "qformer_attention_mask", "pixel_values"] , )
55
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a : Optional[Any] = { 'configuration_gpt_bigcode': ['GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTBigCodeConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ 'GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTBigCodeForSequenceClassification', 'GPTBigCodeForTokenClassification', 'GPTBigCodeForCausalLM', 'GPTBigCodeModel', 'GPTBigCodePreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
56
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black A : Any = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. A : Dict = " def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n" class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): __lowerCAmelCase = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , "models/bert/" ) ) __lowerCAmelCase = self.transformer_dir shutil.copy( os.path.join(__a , "src/transformers/models/bert/modeling_bert.py" ) , os.path.join(self.transformer_dir , "models/bert/modeling_bert.py" ) , ) def snake_case ( self ): __lowerCAmelCase = "src/transformers" shutil.rmtree(self.transformer_dir ) def snake_case ( self , __a , __a , __a , __a=None ): __lowerCAmelCase = comment + f"\nclass {class_name}(nn.Module):\n" + class_code if overwrite_result is not None: __lowerCAmelCase = comment + f"\nclass {class_name}(nn.Module):\n" + overwrite_result __lowerCAmelCase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 ) __lowerCAmelCase = black.format_str(__a , mode=__a ) __lowerCAmelCase = os.path.join(self.transformer_dir , "new_code.py" ) with open(__a , "w" , newline="\n" ) as f: f.write(__a ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(__a ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=__a ) with open(__a , "r" ) as f: self.assertTrue(f.read() , __a ) def snake_case ( self ): __lowerCAmelCase = check_copies.find_code_in_transformers("models.bert.modeling_bert.BertLMPredictionHead" ) self.assertEqual(__a , __a ) def snake_case ( self ): # Base copy consistency self.check_copy_consistency( "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead" , "BertLMPredictionHead" , REFERENCE_CODE + "\n" , ) # With no empty line at the end self.check_copy_consistency( "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead" , "BertLMPredictionHead" , __a , ) # Copy consistency with rename self.check_copy_consistency( "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel" , "TestModelLMPredictionHead" , re.sub("Bert" , "TestModel" , __a ) , ) # Copy consistency with a really long name __lowerCAmelCase = "TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason" self.check_copy_consistency( f"# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}" , f"{long_class_name}LMPredictionHead" , re.sub("Bert" , __a , __a ) , ) # Copy consistency with overwrite self.check_copy_consistency( "# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel" , "TestModelLMPredictionHead" , __a , overwrite_result=re.sub("Bert" , "TestModel" , __a ) , ) def snake_case ( self ): __lowerCAmelCase = check_copies.LOCALIZED_READMES["README_zh-hans.md"] __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the" " Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for" " Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong" " Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1." " **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace)," " released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and" " lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same" " method has been applied to compress GPT2 into" " [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into" " [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation)," " Multilingual BERT into" " [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German" " version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**" " (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders" " as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang" " Luong, Quoc V. Le, Christopher D. Manning." ) __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the" " Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of" " Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian" " Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n" ) __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the" " Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of" " Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian" " Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1." " **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文" " [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and" " lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same" " method has been applied to compress GPT2 into" " [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into" " [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation)," " Multilingual BERT into" " [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German" " version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自" " Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather" " than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le," " Christopher D. Manning 发布。\n" ) __lowerCAmelCase , __lowerCAmelCase = check_copies.convert_to_localized_md( __a , __a , localized_readme["format_model_list"] ) self.assertFalse(__a ) self.assertEqual(__a , __a ) __lowerCAmelCase , __lowerCAmelCase = check_copies.convert_to_localized_md( __a , __a , localized_readme["format_model_list"] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(__a ) __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the" " Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for" " Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong" " Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut." ) __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and" " the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of" " Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian" " Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n" ) __lowerCAmelCase = ( "1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the" " Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of" " Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian" " Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n" ) __lowerCAmelCase , __lowerCAmelCase = check_copies.convert_to_localized_md( __a , __a , localized_readme["format_model_list"] ) # Check if the model link is synchronized. self.assertEqual(__a , __a )
57
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
'''simple docstring''' import sys lowercase_ = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def lowerCamelCase ( __lowerCamelCase : str ) ->int: _SCREAMING_SNAKE_CASE = 1 for digit in s: product *= int(__lowerCamelCase ) return product def lowerCamelCase ( __lowerCamelCase : str = N ) ->int: _SCREAMING_SNAKE_CASE = -sys.maxsize - 1 _SCREAMING_SNAKE_CASE = n[:13] _SCREAMING_SNAKE_CASE = 13 while cur_index < len(__lowerCamelCase ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _SCREAMING_SNAKE_CASE = substr[1:] + n[cur_index] cur_index += 1 else: _SCREAMING_SNAKE_CASE = max(__lowerCamelCase , str_eval(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f"""{solution() = }""")
58
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
from sklearn.metrics import matthews_corrcoef import datasets __lowerCamelCase = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __lowerCamelCase = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __lowerCamelCase = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): def _SCREAMING_SNAKE_CASE (self : str ) -> List[str]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int32" ), "references": datasets.Value("int32" ), } ) , reference_urls=[ "https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html" ] , ) def _SCREAMING_SNAKE_CASE (self : List[Any] , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : List[str]=None ) -> Optional[int]: '''simple docstring''' return { "matthews_correlation": float(matthews_corrcoef(snake_case__ , snake_case__ , sample_weight=snake_case__ ) ), }
59
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowercase : Optional[Any] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") lowercase : Tuple = parser.parse_args() lowercase : Optional[int] = "cpu" lowercase : Optional[Any] = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowercase : Optional[int] = "path-to-your-trained-model" lowercase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowercase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowercase : Dict = pipe.to(device) # to channels last lowercase : Optional[Any] = pipe.unet.to(memory_format=torch.channels_last) lowercase : int = pipe.vae.to(memory_format=torch.channels_last) lowercase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowercase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowercase : Any = torch.randn(2, 4, 64, 64) lowercase : Optional[int] = torch.rand(1) * 999 lowercase : Optional[Any] = torch.randn(2, 77, 768) lowercase : Optional[Any] = (sample, timestep, encoder_hidden_status) try: lowercase : List[Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowercase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Optional[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowercase : Tuple = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowercase : List[str] = 666 lowercase : Tuple = torch.Generator(device).manual_seed(seed) lowercase : Union[str, Any] = {"generator": generator} if args.steps is not None: lowercase : Dict = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowercase : List[str] = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
42
0
"""simple docstring""" def _snake_case ( _snake_case : str , _snake_case : bool = False ): if not isinstance(_snake_case , _snake_case ): lowerCAmelCase : Optional[Any] = f'''Expected string as input, found {type(_snake_case )}''' raise ValueError(_snake_case ) if not isinstance(_snake_case , _snake_case ): lowerCAmelCase : Tuple = f'''Expected boolean as use_pascal parameter, found {type(_snake_case )}''' raise ValueError(_snake_case ) lowerCAmelCase : Any = input_str.split('''_''' ) lowerCAmelCase : str = 0 if use_pascal else 1 lowerCAmelCase : Tuple = words[start_index:] lowerCAmelCase : Dict = [word[0].upper() + word[1:] for word in words_to_capitalize] lowerCAmelCase : Dict = '''''' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
60
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCAmelCase ( _lowerCamelCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return 0.0 def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[int | float, int | float]: _snake_case = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.abs(np.fft.fft(__A ) ) _snake_case = 20 * np.logaa(__A ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds _snake_case = get_bounds(__A , __A ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__A ) plt.show() def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.angle(np.fft.fft(__A ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__A , -2 * pi ) ) plt.show()
42
0
"""simple docstring""" import os from datetime import datetime as dt from github import Github _a = [ 'good first issue', 'feature request', 'wip', ] def __a ( ): UpperCAmelCase_ : int = Github(os.environ["GITHUB_TOKEN"] ) UpperCAmelCase_ : Union[str, Any] = g.get_repo("huggingface/accelerate" ) UpperCAmelCase_ : str = repo.get_issues(state="open" ) for issue in open_issues: UpperCAmelCase_ : Optional[int] = sorted([comment for comment in issue.get_comments()], key=lambda __lowerCamelCase : i.created_at, reverse=__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = comments[0] if len(__lowerCamelCase ) > 0 else None UpperCAmelCase_ : Optional[int] = dt.utcnow() UpperCAmelCase_ : List[Any] = (current_time - issue.updated_at).days UpperCAmelCase_ : Union[str, Any] = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state="closed" ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
61
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class __UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = vocab_size _snake_case = d_embed _snake_case = d_proj _snake_case = cutoffs + [vocab_size] _snake_case = [0] + self.cutoffs _snake_case = div_val _snake_case = self.cutoffs[0] _snake_case = len(self.cutoffs ) - 1 _snake_case = self.shortlist_size + self.n_clusters _snake_case = keep_order _snake_case = [] _snake_case = [] def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if self.n_clusters > 0: _snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_weight' ) _snake_case = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: _snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' , ) self.out_projs.append(lowerCAmelCase_ ) else: self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case = self.d_embed // (self.div_val**i) _snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' ) self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = x if proj is not None: _snake_case = tf.einsum('ibd,ed->ibe' , lowerCAmelCase_ , lowerCAmelCase_ ) return tf.einsum('ibd,nd->ibn' , lowerCAmelCase_ , lowerCAmelCase_ ) + b @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = tf.range(lp_size[0] , dtype=target.dtype ) _snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False ): """simple docstring""" _snake_case = 0 if self.n_clusters == 0: _snake_case = self._logit(lowerCAmelCase_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: _snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase_ , logits=lowerCAmelCase_ ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ , axis=-1 ) else: _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = [] _snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: _snake_case = (target >= l_idx) & (target < r_idx) _snake_case = tf.where(lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) - l_idx if self.div_val == 1: _snake_case = self.out_layers[0][0][l_idx:r_idx] _snake_case = self.out_layers[0][1][l_idx:r_idx] else: _snake_case = self.out_layers[i][0] _snake_case = self.out_layers[i][1] if i == 0: _snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) _snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[0] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) else: _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[i] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) _snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster _snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase_ ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase_ , -cur_logprob , shape_list(lowerCAmelCase_ ) ) _snake_case = tf.concat(lowerCAmelCase_ , axis=-1 ) if target is not None: if return_mean: _snake_case = tf.reduce_mean(lowerCAmelCase_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase_ , name=self.name , aggregation='mean' if return_mean else '' ) return out
42
0
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def _a ( self ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() def _a ( self ) -> List[Any]: __UpperCamelCase , __UpperCamelCase =FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=A_ , dtype=jnp.bfloataa ) __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=A_ , from_pt=A_ , dtype=jnp.bfloataa ) __UpperCamelCase =controlnet_params __UpperCamelCase ='bird' __UpperCamelCase =jax.device_count() __UpperCamelCase =pipe.prepare_text_inputs([prompts] * num_samples ) __UpperCamelCase =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) __UpperCamelCase =pipe.prepare_image_inputs([canny_image] * num_samples ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =jax.random.split(A_ , jax.device_count() ) __UpperCamelCase =replicate(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipe( prompt_ids=A_ , image=A_ , params=A_ , prng_seed=A_ , num_inference_steps=50 , jit=A_ , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __UpperCamelCase =images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __UpperCamelCase =images[0, 253:256, 253:256, -1] __UpperCamelCase =jnp.asarray(jax.device_get(image_slice.flatten() ) ) __UpperCamelCase =jnp.array( [0.16_7969, 0.11_6699, 0.08_1543, 0.15_4297, 0.13_2812, 0.10_8887, 0.16_9922, 0.16_9922, 0.20_5078] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def _a ( self ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase =FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=A_ , dtype=jnp.bfloataa ) __UpperCamelCase , __UpperCamelCase =FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=A_ , from_pt=A_ , dtype=jnp.bfloataa ) __UpperCamelCase =controlnet_params __UpperCamelCase ='Chef in the kitchen' __UpperCamelCase =jax.device_count() __UpperCamelCase =pipe.prepare_text_inputs([prompts] * num_samples ) __UpperCamelCase =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) __UpperCamelCase =pipe.prepare_image_inputs([pose_image] * num_samples ) __UpperCamelCase =jax.random.PRNGKey(0 ) __UpperCamelCase =jax.random.split(A_ , jax.device_count() ) __UpperCamelCase =replicate(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =shard(A_ ) __UpperCamelCase =pipe( prompt_ids=A_ , image=A_ , params=A_ , prng_seed=A_ , num_inference_steps=50 , jit=A_ , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __UpperCamelCase =images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __UpperCamelCase =images[0, 253:256, 253:256, -1] __UpperCamelCase =jnp.asarray(jax.device_get(image_slice.flatten() ) ) __UpperCamelCase =jnp.array( [[0.27_1484, 0.26_1719, 0.27_5391, 0.27_7344, 0.27_9297, 0.29_1016, 0.29_4922, 0.30_2734, 0.30_2734]] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
62
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
0
'''simple docstring''' import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Tuple , __a : List[str] , __a : str=13 , __a : List[str]=7 , __a : Dict=True , __a : Any=True , __a : List[str]=True , __a : List[str]=True , __a : Optional[int]=99 , __a : Union[str, Any]=32 , __a : Optional[Any]=5 , __a : Union[str, Any]=4 , __a : Dict=37 , __a : List[str]="gelu" , __a : Optional[Any]=0.1 , __a : List[Any]=0.1 , __a : Tuple=1_28 , __a : int=32 , __a : int=16 , __a : Dict=2 , __a : Dict=0.02 , __a : Any=3 , __a : Dict=4 , __a : Any=None , ): _a = parent _a = batch_size _a = seq_length _a = is_training _a = use_input_mask _a = use_token_type_ids _a = use_labels _a = vocab_size _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = intermediate_size _a = hidden_act _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = max_position_embeddings _a = type_vocab_size _a = type_sequence_label_size _a = initializer_range _a = num_labels _a = num_choices _a = scope def UpperCamelCase__ ( self : List[Any] ): _a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a = None if self.use_input_mask: _a = random_attention_mask([self.batch_size, self.seq_length] ) _a = None if self.use_token_type_ids: _a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a = None _a = None _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a = ids_tensor([self.batch_size] , self.num_choices ) _a = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self : List[str] ): return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__a , initializer_range=self.initializer_range , ) def UpperCamelCase__ ( self : List[str] ): ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = self.prepare_config_and_inputs() _a = True _a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase__ ( self : Tuple , __a : Optional[Any] , __a : Tuple , __a : Tuple , __a : Union[str, Any] , __a : Tuple , __a : List[Any] , __a : str ): _a = NezhaModel(config=__a ) model.to(__a ) model.eval() _a = model(__a , attention_mask=__a , token_type_ids=__a ) _a = model(__a , token_type_ids=__a ) _a = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self : Optional[Any] , __a : Tuple , __a : List[Any] , __a : Any , __a : List[str] , __a : int , __a : Dict , __a : Optional[int] , __a : Dict , __a : int , ): _a = True _a = NezhaModel(__a ) model.to(__a ) model.eval() _a = model( __a , attention_mask=__a , token_type_ids=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , ) _a = model( __a , attention_mask=__a , token_type_ids=__a , encoder_hidden_states=__a , ) _a = model(__a , attention_mask=__a , token_type_ids=__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self : Optional[Any] , __a : Optional[Any] , __a : Dict , __a : str , __a : Any , __a : Dict , __a : Optional[int] , __a : List[str] ): _a = NezhaForMaskedLM(config=__a ) model.to(__a ) model.eval() _a = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self : List[Any] , __a : Union[str, Any] , __a : str , __a : List[Any] , __a : Optional[int] , __a : Any , __a : Optional[int] , __a : Union[str, Any] ): _a = NezhaForNextSentencePrediction(config=__a ) model.to(__a ) model.eval() _a = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def UpperCamelCase__ ( self : List[str] , __a : Optional[Any] , __a : Union[str, Any] , __a : Any , __a : Any , __a : Optional[int] , __a : Any , __a : Union[str, Any] ): _a = NezhaForPreTraining(config=__a ) model.to(__a ) model.eval() _a = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , next_sentence_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def UpperCamelCase__ ( self : Any , __a : int , __a : Optional[int] , __a : Union[str, Any] , __a : Union[str, Any] , __a : Union[str, Any] , __a : List[str] , __a : List[Any] ): _a = NezhaForQuestionAnswering(config=__a ) model.to(__a ) model.eval() _a = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self : Union[str, Any] , __a : List[str] , __a : List[Any] , __a : Union[str, Any] , __a : Union[str, Any] , __a : List[Any] , __a : List[Any] , __a : Any ): _a = self.num_labels _a = NezhaForSequenceClassification(__a ) model.to(__a ) model.eval() _a = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self : str , __a : str , __a : int , __a : Optional[Any] , __a : Tuple , __a : Tuple , __a : str , __a : List[str] ): _a = self.num_labels _a = NezhaForTokenClassification(config=__a ) model.to(__a ) model.eval() _a = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self : Union[str, Any] , __a : List[str] , __a : str , __a : Optional[int] , __a : int , __a : Any , __a : Dict , __a : Optional[int] ): _a = self.num_choices _a = NezhaForMultipleChoice(config=__a ) model.to(__a ) model.eval() _a = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase__ ( self : Any ): _a = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = config_and_inputs _a = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE (lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __a =( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __a =( { 'feature-extraction': NezhaModel, 'fill-mask': NezhaForMaskedLM, 'question-answering': NezhaForQuestionAnswering, 'text-classification': NezhaForSequenceClassification, 'token-classification': NezhaForTokenClassification, 'zero-shot': NezhaForSequenceClassification, } if is_torch_available() else {} ) __a =True def UpperCamelCase__ ( self : Dict , __a : Union[str, Any] , __a : Any , __a : Optional[int]=False ): _a = super()._prepare_for_class(__a , __a , return_labels=__a ) if return_labels: if model_class in get_values(__a ): _a = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a ) _a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a ) return inputs_dict def UpperCamelCase__ ( self : int ): _a = NezhaModelTester(self ) _a = ConfigTester(self , config_class=__a , hidden_size=37 ) def UpperCamelCase__ ( self : int ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self : Union[str, Any] ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def UpperCamelCase__ ( self : Union[str, Any] ): _a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__a ) def UpperCamelCase__ ( self : str ): # This regression test was failing with PyTorch < 1.3 ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() _a = None self.model_tester.create_and_check_model_as_decoder( __a , __a , __a , __a , __a , __a , __a , __a , __a , ) def UpperCamelCase__ ( self : Optional[int] ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a ) def UpperCamelCase__ ( self : Any ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a ) def UpperCamelCase__ ( self : Tuple ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*__a ) def UpperCamelCase__ ( self : Tuple ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a ) def UpperCamelCase__ ( self : Tuple ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a ) def UpperCamelCase__ ( self : Optional[Any] ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a ) def UpperCamelCase__ ( self : Dict ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__a ) @slow def UpperCamelCase__ ( self : int ): for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = NezhaModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @slow @require_torch_gpu def UpperCamelCase__ ( self : List[str] ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return _a = True _a = model_class(config=__a ) _a = self._prepare_for_class(__a , __a ) _a = torch.jit.trace( __a , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__a , os.path.join(__a , "bert.pt" ) ) _a = torch.jit.load(os.path.join(__a , "bert.pt" ) , map_location=__a ) loaded(inputs_dict["input_ids"].to(__a ) , inputs_dict["attention_mask"].to(__a ) ) @require_torch class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" @slow def UpperCamelCase__ ( self : List[Any] ): _a = NezhaModel.from_pretrained("sijunhe/nezha-cn-base" ) _a = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _a = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _a = model(__a , attention_mask=__a )[0] _a = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , __a ) _a = torch.tensor([[[0.0685, 0.2441, 0.1102], [0.0600, 0.1906, 0.1349], [0.0221, 0.0819, 0.0586]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4 ) ) @slow def UpperCamelCase__ ( self : List[str] ): _a = NezhaForMaskedLM.from_pretrained("sijunhe/nezha-cn-base" ) _a = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _a = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _a = model(__a , attention_mask=__a )[0] _a = torch.Size((1, 6, 2_11_28) ) self.assertEqual(output.shape , __a ) _a = torch.tensor( [[-2.7939, -1.7902, -2.2189], [-2.8585, -1.8908, -2.3723], [-2.6499, -1.7750, -2.2558]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4 ) )
63
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ = 16 A_ = 32 def UpperCAmelCase__ (snake_case__ : Accelerator , snake_case__ : int = 16 ): """simple docstring""" _snake_case : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _snake_case : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ : Any ): # max_length=None => use the model max length (it's actually the default) _snake_case : Any = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : List[Any] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : int = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : Optional[int] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : str = 16 elif accelerator.mixed_precision != "no": _snake_case : Optional[int] = 8 else: _snake_case : Optional[int] = None return tokenizer.pad( snake_case__ , padding="""longest""" , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _snake_case : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) _snake_case : Dict = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , snake_case__ ) == "1": _snake_case : List[Any] = 2 # Initialize accelerator _snake_case : str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Tuple = config["""lr"""] _snake_case : str = int(config["""num_epochs"""] ) _snake_case : Union[str, Any] = int(config["""seed"""] ) _snake_case : Union[str, Any] = int(config["""batch_size"""] ) _snake_case : List[str] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=snake_case__ ) def inner_training_loop(snake_case__ : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : List[Any] = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Tuple = model.to(accelerator.device ) # Instantiate optimizer _snake_case : str = AdamW(params=model.parameters() , lr=snake_case__ ) _snake_case , _snake_case : Optional[int] = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate scheduler _snake_case : str = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=1_00 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case , _snake_case , _snake_case , _snake_case , _snake_case : List[str] = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : int = model(**snake_case__ ) _snake_case : str = outputs.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : int = model(**snake_case__ ) _snake_case : Optional[Any] = outputs.logits.argmax(dim=-1 ) _snake_case , _snake_case : Tuple = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) _snake_case : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , snake_case__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): """simple docstring""" _snake_case : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=snake_case__ , default=snake_case__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _snake_case : Dict = parser.parse_args() _snake_case : int = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
64
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = 1 _snake_case = 2 while i * i <= n: _snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _snake_case = 1 _snake_case = 1 while True: i += 1 t_num += i if count_divisors(__A ) > 500: break return t_num if __name__ == "__main__": print(solution())
42
0
from typing import Any def lowerCAmelCase_ ( __A ) -> list[Any]: '''simple docstring''' if not input_list: return [] UpperCAmelCase__ = [input_list.count(__A ) for value in input_list] UpperCAmelCase__ = max(__A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(__A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
65
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __UpperCAmelCase ( _lowerCamelCase ): @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _snake_case = self.get_env() _snake_case = '1' _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import AutoModel\n ' _snake_case = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
42
0
"""simple docstring""" from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def A_ ( _lowercase = "isbn/0140328726" ): '''simple docstring''' snake_case_ :str = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: snake_case_ :Any = f"""{olid} is not a valid Open Library olid""" raise ValueError(_lowercase ) return requests.get(f"""https://openlibrary.org/{new_olid}.json""" ).json() def A_ ( _lowercase ): '''simple docstring''' snake_case_ :Tuple = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } snake_case_ :Any = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} snake_case_ :Optional[int] = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] snake_case_ :Optional[int] = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(_lowercase, _lowercase ): snake_case_ :Tuple = """, """.join(_lowercase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __a = input("\nEnter the ISBN code to search (or 'quit' to stop): ").strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F"""Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.""") continue print(F"""\nSearching Open Library for ISBN: {isbn}...\n""") try: __a = summarize_book(get_openlibrary_data(F"""isbn/{isbn}""")) print("\n".join(F"""{key}: {value}""" for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F"""Sorry, there are no results for ISBN: {isbn}.""")
66
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCAmelCase_ ).to(lowerCAmelCase_ ) _snake_case = AutoTokenizer.from_pretrained('google/mt5-small' ) _snake_case = tokenizer('Hello there' , return_tensors='pt' ).input_ids _snake_case = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _snake_case = model(input_ids.to(lowerCAmelCase_ ) , labels=labels.to(lowerCAmelCase_ ) ).loss _snake_case = -(labels.shape[-1] * loss.item()) _snake_case = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
42
0
'''simple docstring''' import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated __UpperCAmelCase =collections.namedtuple("_Datasets", ["train", "validation", "test"]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ __UpperCAmelCase ="https://storage.googleapis.com/cvdf-datasets/mnist/" def __lowerCAmelCase ( UpperCamelCase__ ) -> Any: __lowerCamelCase = numpy.dtype(numpy.uintaa ).newbyteorder('''>''' ) return numpy.frombuffer(bytestream.read(4 ) , dtype=UpperCamelCase__ )[0] @deprecated(UpperCamelCase__ , '''Please use tf.data to implement this functionality.''' ) def __lowerCAmelCase ( UpperCamelCase__ ) -> Optional[Any]: print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=UpperCamelCase__ ) as bytestream: __lowerCamelCase = _readaa(UpperCamelCase__ ) if magic != 20_51: raise ValueError( '''Invalid magic number %d in MNIST image file: %s''' % (magic, f.name) ) __lowerCamelCase = _readaa(UpperCamelCase__ ) __lowerCamelCase = _readaa(UpperCamelCase__ ) __lowerCamelCase = _readaa(UpperCamelCase__ ) __lowerCamelCase = bytestream.read(rows * cols * num_images ) __lowerCamelCase = numpy.frombuffer(UpperCamelCase__ , dtype=numpy.uinta ) __lowerCamelCase = data.reshape(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , 1 ) return data @deprecated(UpperCamelCase__ , '''Please use tf.one_hot on tensors.''' ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Dict: __lowerCamelCase = labels_dense.shape[0] __lowerCamelCase = numpy.arange(UpperCamelCase__ ) * num_classes __lowerCamelCase = numpy.zeros((num_labels, num_classes) ) __lowerCamelCase = 1 return labels_one_hot @deprecated(UpperCamelCase__ , '''Please use tf.data to implement this functionality.''' ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__=10 ) -> Union[str, Any]: print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=UpperCamelCase__ ) as bytestream: __lowerCamelCase = _readaa(UpperCamelCase__ ) if magic != 20_49: raise ValueError( '''Invalid magic number %d in MNIST label file: %s''' % (magic, f.name) ) __lowerCamelCase = _readaa(UpperCamelCase__ ) __lowerCamelCase = bytestream.read(UpperCamelCase__ ) __lowerCamelCase = numpy.frombuffer(UpperCamelCase__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(UpperCamelCase__ , UpperCamelCase__ ) return labels class a__ : @deprecated( a , '''Please use alternatives such as official/mnist/_DataSet.py''' ''' from tensorflow/models.''' , ) def __init__( self : List[Any] , a : List[Any] , a : Tuple , a : Tuple=False , a : Tuple=False , a : Any=dtypes.floataa , a : Union[str, Any]=True , a : Tuple=None , ): """simple docstring""" __lowerCamelCase , __lowerCamelCase = random_seed.get_seed(a ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) __lowerCamelCase = dtypes.as_dtype(a ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError('''Invalid image dtype %r, expected uint8 or float32''' % dtype ) if fake_data: __lowerCamelCase = 1_00_00 __lowerCamelCase = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f"""images.shape: {images.shape} labels.shape: {labels.shape}""" __lowerCamelCase = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 __lowerCamelCase = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. __lowerCamelCase = images.astype(numpy.floataa ) __lowerCamelCase = numpy.multiply(a , 1.0 / 2_55.0 ) __lowerCamelCase = images __lowerCamelCase = labels __lowerCamelCase = 0 __lowerCamelCase = 0 @property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return self._images @property def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" return self._labels @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return self._num_examples @property def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" return self._epochs_completed def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , a : Any , a : List[Any]=False , a : Union[str, Any]=True ): """simple docstring""" if fake_data: __lowerCamelCase = [1] * 7_84 __lowerCamelCase = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(a )], [fake_label for _ in range(a )], ) __lowerCamelCase = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: __lowerCamelCase = numpy.arange(self._num_examples ) numpy.random.shuffle(a ) __lowerCamelCase = self.images[perma] __lowerCamelCase = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch __lowerCamelCase = self._num_examples - start __lowerCamelCase = self._images[start : self._num_examples] __lowerCamelCase = self._labels[start : self._num_examples] # Shuffle the data if shuffle: __lowerCamelCase = numpy.arange(self._num_examples ) numpy.random.shuffle(a ) __lowerCamelCase = self.images[perm] __lowerCamelCase = self.labels[perm] # Start next epoch __lowerCamelCase = 0 __lowerCamelCase = batch_size - rest_num_examples __lowerCamelCase = self._index_in_epoch __lowerCamelCase = self._images[start:end] __lowerCamelCase = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size __lowerCamelCase = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(UpperCamelCase__ , '''Please write your own downloading logic.''' ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: if not gfile.Exists(UpperCamelCase__ ): gfile.MakeDirs(UpperCamelCase__ ) __lowerCamelCase = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) if not gfile.Exists(UpperCamelCase__ ): urllib.request.urlretrieve(UpperCamelCase__ , UpperCamelCase__ ) # noqa: S310 with gfile.GFile(UpperCamelCase__ ) as f: __lowerCamelCase = f.size() print('''Successfully downloaded''' , UpperCamelCase__ , UpperCamelCase__ , '''bytes.''' ) return filepath @deprecated( UpperCamelCase__ , '''Please use alternatives such as:''' ''' tensorflow_datasets.load(\'mnist\')''' ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__=False , UpperCamelCase__=dtypes.floataa , UpperCamelCase__=True , UpperCamelCase__=50_00 , UpperCamelCase__=None , UpperCamelCase__=DEFAULT_SOURCE_URL , ) -> Tuple: if fake_data: def fake(): return _DataSet( [] , [] , fake_data=UpperCamelCase__ , one_hot=UpperCamelCase__ , dtype=UpperCamelCase__ , seed=UpperCamelCase__ ) __lowerCamelCase = fake() __lowerCamelCase = fake() __lowerCamelCase = fake() return _Datasets(train=UpperCamelCase__ , validation=UpperCamelCase__ , test=UpperCamelCase__ ) if not source_url: # empty string check __lowerCamelCase = DEFAULT_SOURCE_URL __lowerCamelCase = '''train-images-idx3-ubyte.gz''' __lowerCamelCase = '''train-labels-idx1-ubyte.gz''' __lowerCamelCase = '''t10k-images-idx3-ubyte.gz''' __lowerCamelCase = '''t10k-labels-idx1-ubyte.gz''' __lowerCamelCase = _maybe_download( UpperCamelCase__ , UpperCamelCase__ , source_url + train_images_file ) with gfile.Open(UpperCamelCase__ , '''rb''' ) as f: __lowerCamelCase = _extract_images(UpperCamelCase__ ) __lowerCamelCase = _maybe_download( UpperCamelCase__ , UpperCamelCase__ , source_url + train_labels_file ) with gfile.Open(UpperCamelCase__ , '''rb''' ) as f: __lowerCamelCase = _extract_labels(UpperCamelCase__ , one_hot=UpperCamelCase__ ) __lowerCamelCase = _maybe_download( UpperCamelCase__ , UpperCamelCase__ , source_url + test_images_file ) with gfile.Open(UpperCamelCase__ , '''rb''' ) as f: __lowerCamelCase = _extract_images(UpperCamelCase__ ) __lowerCamelCase = _maybe_download( UpperCamelCase__ , UpperCamelCase__ , source_url + test_labels_file ) with gfile.Open(UpperCamelCase__ , '''rb''' ) as f: __lowerCamelCase = _extract_labels(UpperCamelCase__ , one_hot=UpperCamelCase__ ) if not 0 <= validation_size <= len(UpperCamelCase__ ): __lowerCamelCase = ( '''Validation size should be between 0 and ''' f"""{len(UpperCamelCase__ )}. Received: {validation_size}.""" ) raise ValueError(UpperCamelCase__ ) __lowerCamelCase = train_images[:validation_size] __lowerCamelCase = train_labels[:validation_size] __lowerCamelCase = train_images[validation_size:] __lowerCamelCase = train_labels[validation_size:] __lowerCamelCase = {'''dtype''': dtype, '''reshape''': reshape, '''seed''': seed} __lowerCamelCase = _DataSet(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) __lowerCamelCase = _DataSet(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) __lowerCamelCase = _DataSet(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) return _Datasets(train=UpperCamelCase__ , validation=UpperCamelCase__ , test=UpperCamelCase__ )
67
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[str] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: list[int] ) -> int: '''simple docstring''' if not numbers: return 0 if not isinstance(SCREAMING_SNAKE_CASE_ , (list, tuple) ) or not all( isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for number in numbers ): raise ValueError("numbers must be an iterable of integers" ) A__ = A__ = A__ = numbers[0] for i in range(1 , len(SCREAMING_SNAKE_CASE_ ) ): # update the maximum and minimum subarray products A__ = numbers[i] if number < 0: A__ , A__ = min_till_now, max_till_now A__ = max(SCREAMING_SNAKE_CASE_ , max_till_now * number ) A__ = min(SCREAMING_SNAKE_CASE_ , min_till_now * number ) # update the maximum product found till now A__ = max(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return max_prod
68
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowercase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowercase : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowercase : set[int] = {ord(char) for char in VALID_CHARS} lowercase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str | None: _snake_case = "" _snake_case = 42 _snake_case = 42 _snake_case = 42 for keychar, cipherchar in zip(cycle(__A ) , __A ): _snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__A ) return decoded def SCREAMING_SNAKE_CASE__ ( __A ) -> list[str]: _snake_case = [] for key in product(__A , repeat=3 ): _snake_case = try_key(__A , __A ) if encoded is not None: possibles.append(__A ) return possibles def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __A = "p059_cipher.txt" ) -> int: _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = Path(__A ).parent.joinpath(__A ).read_text(encoding='utf-8' ) _snake_case = [int(__A ) for number in data.strip().split(',' )] _snake_case = filter_valid_chars(__A ) for common_word in COMMON_WORDS: _snake_case = filter_common_word(__A , __A ) if len(__A ) == 1: break _snake_case = possibles[0] return sum(ord(__A ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' import warnings from ..trainer import Trainer from ..utils import logging A__ : List[Any] =logging.get_logger(__name__) class UpperCAmelCase ( snake_case_ ): def __init__( self : Tuple , __snake_case : int=None , **__snake_case : Any ) -> Dict: warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , __snake_case , ) super().__init__(args=__snake_case , **__snake_case )
70
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer A_ :Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ :Dict = { '''vocab_file''': { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''unc-nlp/lxmert-base-uncased''': ( '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ :str = { '''unc-nlp/lxmert-base-uncased''': 512, } A_ :Dict = { '''unc-nlp/lxmert-base-uncased''': {'''do_lower_case''': True}, } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[Any] =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Union[str, Any] =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : str =LxmertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : Dict =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCamelCase__ ) != tokenize_chinese_chars ): __UpperCamelCase : str =getattr(lowerCamelCase__ , normalizer_state.pop('type' ) ) __UpperCamelCase : Any =do_lower_case __UpperCamelCase : Dict =strip_accents __UpperCamelCase : List[str] =tokenize_chinese_chars __UpperCamelCase : Union[str, Any] =normalizer_class(**lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =do_lower_case def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" __UpperCamelCase : List[str] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : str =[self.sep_token_id] __UpperCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : int =self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
71
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" lowerCAmelCase__ = [0, 2, 4, 6, 8] lowerCAmelCase__ = [1, 3, 5, 7, 9] def snake_case_ ( A_ : int, A_ : int, A_ : list[int], A_ : int ): '''simple docstring''' if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1, -1, -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 _lowerCamelCase : int = 0 for digit in range(10 ): _lowerCamelCase : List[str] = digit result += reversible_numbers( 0, (remainder + 2 * digit) // 10, A_, A_ ) return result _lowerCamelCase : List[Any] = 0 for digita in range(10 ): _lowerCamelCase : List[str] = digita if (remainder + digita) % 2 == 0: _lowerCamelCase : Tuple = ODD_DIGITS else: _lowerCamelCase : List[str] = EVEN_DIGITS for digita in other_parity_digits: _lowerCamelCase : int = digita result += reversible_numbers( remaining_length - 2, (remainder + digita + digita) // 10, A_, A_, ) return result def snake_case_ ( A_ : int = 9 ): '''simple docstring''' _lowerCamelCase : Optional[Any] = 0 for length in range(1, max_power + 1 ): result += reversible_numbers(A_, 0, [0] * length, A_ ) return result if __name__ == "__main__": print(F"""{solution() = }""")
72
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
import string from math import logaa def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> int: __lowerCamelCase : Tuple = document.translate( str.maketrans('' , '' , string.punctuation ) ).replace('\n' , '' ) __lowerCamelCase : int = document_without_punctuation.split(' ' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> tuple[int, int]: __lowerCamelCase : List[Any] = corpus.lower().translate( str.maketrans('' , '' , string.punctuation ) ) # strip all punctuation and replace it with '' __lowerCamelCase : Union[str, Any] = corpus_without_punctuation.split('\n' ) __lowerCamelCase : Optional[int] = term.lower() return (len([doc for doc in docs if term in doc] ), len(lowerCamelCase__ )) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ) -> float: if smoothing: if n == 0: raise ValueError('log10(0) is undefined.' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('df must be > 0' ) elif n == 0: raise ValueError('log10(0) is undefined.' ) return round(logaa(n / df ) , 3 ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> float: return round(tf * idf , 3 )
73
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: List[str] = '''gpt_bigcode''' _lowerCamelCase: List[Any] = ['''past_key_values'''] _lowerCamelCase: int = { '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[int] ,A_ : Dict=5_0257 ,A_ : Union[str, Any]=1024 ,A_ : str=768 ,A_ : Any=12 ,A_ : Any=12 ,A_ : Optional[int]=None ,A_ : Any="gelu_pytorch_tanh" ,A_ : List[str]=0.1 ,A_ : Optional[int]=0.1 ,A_ : List[str]=0.1 ,A_ : Tuple=1e-5 ,A_ : Optional[int]=0.02 ,A_ : List[str]=True ,A_ : Optional[Any]=True ,A_ : List[Any]=5_0256 ,A_ : Union[str, Any]=5_0256 ,A_ : int=True ,A_ : Optional[Any]=True ,A_ : Dict=True ,**A_ : Union[str, Any] ,) -> Union[str, Any]: A = vocab_size A = n_positions A = n_embd A = n_layer A = n_head A = n_inner A = activation_function A = resid_pdrop A = embd_pdrop A = attn_pdrop A = layer_norm_epsilon A = initializer_range A = scale_attn_weights A = use_cache A = attention_softmax_in_fpaa A = scale_attention_softmax_in_fpaa A = multi_query A = bos_token_id A = eos_token_id super().__init__(bos_token_id=A_ ,eos_token_id=A_ ,**A_ )
74
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
'''simple docstring''' import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def a_ ( __snake_case : Any ) -> int: """simple docstring""" lowerCamelCase_ =checkpoints.load_tax_checkpoint(__snake_case ) lowerCamelCase_ =flatten_dict(__snake_case ) return flax_params def a_ ( __snake_case : Dict ) -> Optional[int]: """simple docstring""" lowerCamelCase_ ={} lowerCamelCase_ ={ '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } lowerCamelCase_ ={ '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowerCamelCase_ ='''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowerCamelCase_ =new_key.replace(__snake_case , __snake_case ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowerCamelCase_ =new_key.replace(__snake_case , __snake_case ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowerCamelCase_ =re.sub(r'''layers_(\d+)''' , r'''layer.\1''' , __snake_case ) lowerCamelCase_ =new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowerCamelCase_ =re.sub(r'''layers_(\d+)''' , r'''layer.\1''' , __snake_case ) lowerCamelCase_ =flax_dict[key] lowerCamelCase_ ={} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowerCamelCase_ =torch.from_numpy(converted_dict[key].T ) else: lowerCamelCase_ =torch.from_numpy(converted_dict[key] ) return converted_torch_dict def a_ ( __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Any=False , __snake_case : Optional[int]=False ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ =get_flax_param(__snake_case ) if not use_large: lowerCamelCase_ =PixaStructVisionConfig() lowerCamelCase_ =PixaStructTextConfig() else: lowerCamelCase_ =PixaStructVisionConfig( hidden_size=1536 , d_ff=3968 , num_attention_heads=24 , num_hidden_layers=18 ) lowerCamelCase_ =PixaStructTextConfig(hidden_size=1536 , d_ff=3968 , num_heads=24 , num_layers=18 ) lowerCamelCase_ =PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__snake_case ) lowerCamelCase_ =PixaStructForConditionalGeneration(__snake_case ) lowerCamelCase_ =rename_and_convert_flax_params(__snake_case ) model.load_state_dict(__snake_case ) lowerCamelCase_ =AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) lowerCamelCase_ =PixaStructImageProcessor() lowerCamelCase_ =PixaStructProcessor(image_processor=__snake_case , tokenizer=__snake_case ) if use_large: lowerCamelCase_ =4096 lowerCamelCase_ =True # mkdir if needed os.makedirs(__snake_case , exist_ok=__snake_case ) model.save_pretrained(__snake_case ) processor.save_pretrained(__snake_case ) print('''Model saved in {}'''.format(__snake_case ) ) if __name__ == "__main__": a_ : Optional[int] = argparse.ArgumentParser() parser.add_argument("""--t5x_checkpoint_path""", default=None, type=str, help="""Path to the original T5x checkpoint.""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--use_large""", action="""store_true""", help="""Use large model.""") parser.add_argument("""--is_vqa""", action="""store_true""", help="""Use large model.""") a_ : Tuple = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
75
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor a_ = logging.get_logger(__name__) class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : int , *a : List[Any] , **a : Any ) -> None: """simple docstring""" warnings.warn( "The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use DeformableDetrImageProcessor instead." , a , ) super().__init__(*a , **a )
76
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
"""simple docstring""" from __future__ import annotations def a_ ( _lowerCAmelCase : int | float | str , _lowerCAmelCase : int | float | str ): '''simple docstring''' if nth_term == "": return [""] lowercase__ : List[Any] = int(_lowerCAmelCase ) lowercase__ : List[Any] = int(_lowerCAmelCase ) lowercase__ : list[str] = [] for temp in range(int(_lowerCAmelCase ) ): series.append(f"""1 / {pow(temp + 1 , int(_lowerCAmelCase ) )}""" if series else '1' ) return series if __name__ == "__main__": import doctest doctest.testmod() _UpperCamelCase : Tuple = int(input("Enter the last number (nth term) of the P-Series")) _UpperCamelCase : str = int(input("Enter the power for P-Series")) print("Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p") print(p_series(nth_term, power))
77
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case_ = { """configuration_bloom""": ["""BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BloomConfig""", """BloomOnnxConfig"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = ["""BloomTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ """BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST""", """BloomForCausalLM""", """BloomModel""", """BloomPreTrainedModel""", """BloomForSequenceClassification""", """BloomForTokenClassification""", """BloomForQuestionAnswering""", ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0
'''simple docstring''' def __lowercase ( __lowercase , __lowercase , __lowercase=False ) -> Union[str, Any]: '''simple docstring''' if isinstance(__lowercase , __lowercase ) and isinstance(__lowercase , __lowercase ): _A = len(set_a.intersection(__lowercase ) ) if alternative_union: _A = len(__lowercase ) + len(__lowercase ) else: _A = len(set_a.union(__lowercase ) ) return intersection / union if isinstance(__lowercase , (list, tuple) ) and isinstance(__lowercase , (list, tuple) ): _A = [element for element in set_a if element in set_b] if alternative_union: _A = len(__lowercase ) + len(__lowercase ) return len(__lowercase ) / union else: _A = set_a + [element for element in set_b if element not in set_a] return len(__lowercase ) / len(__lowercase ) return len(__lowercase ) / len(__lowercase ) return None if __name__ == "__main__": lowerCamelCase_ = {'''a''', '''b''', '''c''', '''d''', '''e'''} lowerCamelCase_ = {'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
79
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
'''simple docstring''' import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _UpperCamelCase ( __A , __A , __A , __A , __A ) -> List[str]: '''simple docstring''' with open(__A ) as metadata_file: UpperCamelCase__ = json.load(__A ) UpperCamelCase__ = LukeConfig(use_entity_aware_attention=__A , **metadata["model_config"] ) # Load in the weights from the checkpoint_path UpperCamelCase__ = torch.load(__A , map_location="cpu" )["module"] # Load the entity vocab file UpperCamelCase__ = load_original_entity_vocab(__A ) # add an entry for [MASK2] UpperCamelCase__ = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 UpperCamelCase__ = XLMRobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks UpperCamelCase__ = AddedToken("<ent>" , lstrip=__A , rstrip=__A ) UpperCamelCase__ = AddedToken("<ent2>" , lstrip=__A , rstrip=__A ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__A ) with open(os.path.join(__A , "tokenizer_config.json" ) , "r" ) as f: UpperCamelCase__ = json.load(__A ) UpperCamelCase__ = "MLukeTokenizer" with open(os.path.join(__A , "tokenizer_config.json" ) , "w" ) as f: json.dump(__A , __A ) with open(os.path.join(__A , MLukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(__A , __A ) UpperCamelCase__ = MLukeTokenizer.from_pretrained(__A ) # Initialize the embeddings of the special tokens UpperCamelCase__ = tokenizer.convert_tokens_to_ids(["@"] )[0] UpperCamelCase__ = tokenizer.convert_tokens_to_ids(["#"] )[0] UpperCamelCase__ = state_dict["embeddings.word_embeddings.weight"] UpperCamelCase__ = word_emb[ent_init_index].unsqueeze(0 ) UpperCamelCase__ = word_emb[enta_init_index].unsqueeze(0 ) UpperCamelCase__ = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: UpperCamelCase__ = state_dict[bias_name] UpperCamelCase__ = decoder_bias[ent_init_index].unsqueeze(0 ) UpperCamelCase__ = decoder_bias[enta_init_index].unsqueeze(0 ) UpperCamelCase__ = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: UpperCamelCase__ = F'''encoder.layer.{layer_index}.attention.self.''' UpperCamelCase__ = state_dict[prefix + matrix_name] UpperCamelCase__ = state_dict[prefix + matrix_name] UpperCamelCase__ = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks UpperCamelCase__ = state_dict["entity_embeddings.entity_embeddings.weight"] UpperCamelCase__ = entity_emb[entity_vocab["[MASK]"]].unsqueeze(0 ) UpperCamelCase__ = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' UpperCamelCase__ = state_dict["entity_predictions.bias"] UpperCamelCase__ = entity_prediction_bias[entity_vocab["[MASK]"]].unsqueeze(0 ) UpperCamelCase__ = torch.cat([entity_prediction_bias, entity_mask_bias] ) UpperCamelCase__ = LukeForMaskedLM(config=__A ).eval() state_dict.pop("entity_predictions.decoder.weight" ) state_dict.pop("lm_head.decoder.weight" ) state_dict.pop("lm_head.decoder.bias" ) UpperCamelCase__ = OrderedDict() for key, value in state_dict.items(): if not (key.startswith("lm_head" ) or key.startswith("entity_predictions" )): UpperCamelCase__ = state_dict[key] else: UpperCamelCase__ = state_dict[key] UpperCamelCase__ , UpperCamelCase__ = model.load_state_dict(__A , strict=__A ) if set(__A ) != {"luke.embeddings.position_ids"}: raise ValueError(F'''Unexpected unexpected_keys: {unexpected_keys}''' ) if set(__A ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F'''Unexpected missing_keys: {missing_keys}''' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs UpperCamelCase__ = MLukeTokenizer.from_pretrained(__A , task="entity_classification" ) UpperCamelCase__ = "ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan)." UpperCamelCase__ = (0, 9) UpperCamelCase__ = tokenizer(__A , entity_spans=[span] , return_tensors="pt" ) UpperCamelCase__ = model(**__A ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base UpperCamelCase__ = torch.Size((1, 33, 768) ) UpperCamelCase__ = torch.tensor([[0.0892, 0.0596, -0.2819], [0.0134, 0.1199, 0.0573], [-0.0169, 0.0927, 0.0644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __A , atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base UpperCamelCase__ = torch.Size((1, 1, 768) ) UpperCamelCase__ = torch.tensor([[-0.1482, 0.0609, 0.0322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __A , atol=1E-4 ): raise ValueError # Verify masked word/entity prediction UpperCamelCase__ = MLukeTokenizer.from_pretrained(__A ) UpperCamelCase__ = "Tokyo is the capital of <mask>." UpperCamelCase__ = (24, 30) UpperCamelCase__ = tokenizer(__A , entity_spans=[span] , return_tensors="pt" ) UpperCamelCase__ = model(**__A ) UpperCamelCase__ = encoding["input_ids"][0].tolist() UpperCamelCase__ = input_ids.index(tokenizer.convert_tokens_to_ids("<mask>" ) ) UpperCamelCase__ = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(__A ) UpperCamelCase__ = outputs.entity_logits[0][0].argmax().item() UpperCamelCase__ = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith("en:" )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(__A ) ) model.save_pretrained(__A ) def _UpperCamelCase ( __A ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = ["[MASK]", "[PAD]", "[UNK]"] UpperCamelCase__ = [json.loads(__A ) for line in open(__A )] UpperCamelCase__ = {} for entry in data: UpperCamelCase__ = entry["id"] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: UpperCamelCase__ = entity_id break UpperCamelCase__ = F'''{language}:{entity_name}''' UpperCamelCase__ = entity_id return new_mapping if __name__ == "__main__": a__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Path to a pytorch_model.bin file.') parser.add_argument( '--metadata_path', default=None, type=str, help='Path to a metadata.json file, defining the configuration.' ) parser.add_argument( '--entity_vocab_path', default=None, type=str, help='Path to an entity_vocab.tsv file, containing the entity vocabulary.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to where to dump the output PyTorch model.' ) parser.add_argument( '--model_size', default='base', type=str, choices=['base', 'large'], help='Size of the model to be converted.' ) a__ : Any = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
80
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, is_batched, to_numpy_array, valid_images, ) from ...utils import TensorType, logging lowerCamelCase_ : List[Any] = logging.get_logger(__name__) class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = ["pixel_values"] def __init__( self , __A = True , __A = None , __A = PILImageResampling.BICUBIC , __A = True , __A = True , __A = 1 / 255 , __A = None , __A = True , __A = None , __A = None , **__A , ) -> None: super().__init__(**__A ) a =size if size is not None else {'''height''': 224, '''width''': 224} a =get_size_dict(__A ) a =crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} a =get_size_dict(__A , default_to_square=__A , param_name='''crop_size''' ) a =do_resize a =do_rescale a =do_normalize a =do_center_crop a =crop_size a =size a =resample a =rescale_factor a =image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN a =image_std if image_std is not None else IMAGENET_DEFAULT_STD def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = PILImageResampling.BILINEAR , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A ) if "shortest_edge" in size: a =get_resize_output_image_size(__A , size=size['''shortest_edge'''] , default_to_square=__A ) # size = get_resize_output_image_size(image, size["shortest_edge"], size["longest_edge"]) elif "height" in size and "width" in size: a =(size['''height'''], size['''width''']) else: raise ValueError(f'''Size must contain \'height\' and \'width\' keys or \'shortest_edge\' key. Got {size.keys()}''' ) return resize(__A , size=__A , resample=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(__A , size=(size['''height'''], size['''width''']) , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A ) -> np.ndarray: return rescale(__A , scale=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A = None , **__A , ) -> np.ndarray: return normalize(__A , mean=__A , std=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = ChannelDimension.FIRST , **__A , ) -> BatchFeature: a =do_resize if do_resize is not None else self.do_resize a =do_rescale if do_rescale is not None else self.do_rescale a =do_normalize if do_normalize is not None else self.do_normalize a =do_center_crop if do_center_crop is not None else self.do_center_crop a =crop_size if crop_size is not None else self.crop_size a =get_size_dict(__A , param_name='''crop_size''' , default_to_square=__A ) a =resample if resample is not None else self.resample a =rescale_factor if rescale_factor is not None else self.rescale_factor a =image_mean if image_mean is not None else self.image_mean a =image_std if image_std is not None else self.image_std a =size if size is not None else self.size a =get_size_dict(__A ) if not is_batched(__A ): a =[images] if not valid_images(__A ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. a =[to_numpy_array(__A ) for image in images] if do_resize: a =[self.resize(image=__A , size=__A , resample=__A ) for image in images] if do_center_crop: a =[self.center_crop(image=__A , size=__A ) for image in images] if do_rescale: a =[self.rescale(image=__A , scale=__A ) for image in images] if do_normalize: a =[self.normalize(image=__A , mean=__A , std=__A ) for image in images] a =[to_channel_dimension_format(__A , __A ) for image in images] a ={'''pixel_values''': images} return BatchFeature(data=__A , tensor_type=__A )
81
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowercase : Optional[Any] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") lowercase : Tuple = parser.parse_args() lowercase : Optional[int] = "cpu" lowercase : Optional[Any] = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowercase : Optional[int] = "path-to-your-trained-model" lowercase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowercase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowercase : Dict = pipe.to(device) # to channels last lowercase : Optional[Any] = pipe.unet.to(memory_format=torch.channels_last) lowercase : int = pipe.vae.to(memory_format=torch.channels_last) lowercase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowercase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowercase : Any = torch.randn(2, 4, 64, 64) lowercase : Optional[int] = torch.rand(1) * 999 lowercase : Optional[Any] = torch.randn(2, 77, 768) lowercase : Optional[Any] = (sample, timestep, encoder_hidden_status) try: lowercase : List[Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowercase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Optional[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowercase : Tuple = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowercase : List[str] = 666 lowercase : Tuple = torch.Generator(device).manual_seed(seed) lowercase : Union[str, Any] = {"generator": generator} if args.steps is not None: lowercase : Dict = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowercase : List[str] = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
42
0
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent A__ = {"""UserAgent""": UserAgent().random} def _UpperCAmelCase ( snake_case ): """simple docstring""" _lowerCAmelCase = script.contents[0] _lowerCAmelCase = json.loads(data[data.find("""{\"config\"""" ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class __lowerCAmelCase : def __init__( self , _snake_case ): """simple docstring""" _lowerCAmelCase = F'https://www.instagram.com/{username}/' _lowerCAmelCase = self.get_json() def snake_case ( self ): """simple docstring""" _lowerCAmelCase = requests.get(self.url , headers=_snake_case ).text _lowerCAmelCase = BeautifulSoup(_snake_case , """html.parser""" ).find_all("""script""" ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self ): """simple docstring""" return F'{self.__class__.__name__}(\'{self.username}\')' def __str__( self ): """simple docstring""" return F'{self.fullname} ({self.username}) is {self.biography}' @property def snake_case ( self ): """simple docstring""" return self.user_data["username"] @property def snake_case ( self ): """simple docstring""" return self.user_data["full_name"] @property def snake_case ( self ): """simple docstring""" return self.user_data["biography"] @property def snake_case ( self ): """simple docstring""" return self.user_data["business_email"] @property def snake_case ( self ): """simple docstring""" return self.user_data["external_url"] @property def snake_case ( self ): """simple docstring""" return self.user_data["edge_followed_by"]["count"] @property def snake_case ( self ): """simple docstring""" return self.user_data["edge_follow"]["count"] @property def snake_case ( self ): """simple docstring""" return self.user_data["edge_owner_to_timeline_media"]["count"] @property def snake_case ( self ): """simple docstring""" return self.user_data["profile_pic_url_hd"] @property def snake_case ( self ): """simple docstring""" return self.user_data["is_verified"] @property def snake_case ( self ): """simple docstring""" return self.user_data["is_private"] def _UpperCAmelCase ( snake_case = "github" ): """simple docstring""" import os if os.environ.get("""CI""" ): return # test failing on GitHub Actions _lowerCAmelCase = InstagramUser(snake_case ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , snake_case ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 1_50 assert instagram_user.number_of_followers > 12_00_00 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith("""https://instagram.""" ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() A__ = InstagramUser("""github""") print(instagram_user) print(f"{instagram_user.number_of_posts = }") print(f"{instagram_user.number_of_followers = }") print(f"{instagram_user.number_of_followings = }") print(f"{instagram_user.email = }") print(f"{instagram_user.website = }") print(f"{instagram_user.profile_picture_url = }") print(f"{instagram_user.is_verified = }") print(f"{instagram_user.is_private = }")
82
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCAmelCase ( _lowerCamelCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return 0.0 def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[int | float, int | float]: _snake_case = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.abs(np.fft.fft(__A ) ) _snake_case = 20 * np.logaa(__A ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds _snake_case = get_bounds(__A , __A ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__A ) plt.show() def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.angle(np.fft.fft(__A ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__A , -2 * pi ) ) plt.show()
42
0
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class __UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = vocab_size _snake_case = d_embed _snake_case = d_proj _snake_case = cutoffs + [vocab_size] _snake_case = [0] + self.cutoffs _snake_case = div_val _snake_case = self.cutoffs[0] _snake_case = len(self.cutoffs ) - 1 _snake_case = self.shortlist_size + self.n_clusters _snake_case = keep_order _snake_case = [] _snake_case = [] def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if self.n_clusters > 0: _snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_weight' ) _snake_case = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: _snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' , ) self.out_projs.append(lowerCAmelCase_ ) else: self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case = self.d_embed // (self.div_val**i) _snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' ) self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = x if proj is not None: _snake_case = tf.einsum('ibd,ed->ibe' , lowerCAmelCase_ , lowerCAmelCase_ ) return tf.einsum('ibd,nd->ibn' , lowerCAmelCase_ , lowerCAmelCase_ ) + b @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = tf.range(lp_size[0] , dtype=target.dtype ) _snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False ): """simple docstring""" _snake_case = 0 if self.n_clusters == 0: _snake_case = self._logit(lowerCAmelCase_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: _snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase_ , logits=lowerCAmelCase_ ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ , axis=-1 ) else: _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = [] _snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: _snake_case = (target >= l_idx) & (target < r_idx) _snake_case = tf.where(lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) - l_idx if self.div_val == 1: _snake_case = self.out_layers[0][0][l_idx:r_idx] _snake_case = self.out_layers[0][1][l_idx:r_idx] else: _snake_case = self.out_layers[i][0] _snake_case = self.out_layers[i][1] if i == 0: _snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) _snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[0] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) else: _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[i] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) _snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster _snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase_ ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase_ , -cur_logprob , shape_list(lowerCAmelCase_ ) ) _snake_case = tf.concat(lowerCAmelCase_ , axis=-1 ) if target is not None: if return_mean: _snake_case = tf.reduce_mean(lowerCAmelCase_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase_ , name=self.name , aggregation='mean' if return_mean else '' ) return out
42
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def _snake_case ( lowercase__ : Tuple ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ :Union[str, Any] = args.pruning_method lowerCAmelCase_ :Tuple = args.threshold lowerCAmelCase_ :int = args.model_name_or_path.rstrip("""/""" ) lowerCAmelCase_ :Dict = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase_ :Optional[int] = torch.load(os.path.join(lowercase__ , """pytorch_model.bin""" ) ) lowerCAmelCase_ :Union[str, Any] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase_ :Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase_ :Any = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase_ :List[str] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase_ :Any = MagnitudeBinarizer.apply(inputs=lowercase__ , threshold=lowercase__ ) lowerCAmelCase_ :str = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase_ :str = name[:-6] lowerCAmelCase_ :Any = model[f"""{prefix_}mask_scores"""] lowerCAmelCase_ :Optional[Any] = TopKBinarizer.apply(lowercase__ , lowercase__ ) lowerCAmelCase_ :int = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase_ :int = name[:-6] lowerCAmelCase_ :Optional[Any] = model[f"""{prefix_}mask_scores"""] lowerCAmelCase_ :Union[str, Any] = ThresholdBinarizer.apply(lowercase__ , lowercase__ , lowercase__ ) lowerCAmelCase_ :Optional[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase_ :Dict = name[:-6] lowerCAmelCase_ :Any = model[f"""{prefix_}mask_scores"""] lowerCAmelCase_ , lowerCAmelCase_ :str = -0.1, 1.1 lowerCAmelCase_ :Optional[int] = torch.sigmoid(lowercase__ ) lowerCAmelCase_ :Any = s * (r - l) + l lowerCAmelCase_ :Dict = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase_ :Any = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError("""Unknown pruning method""" ) if target_model_path is None: lowerCAmelCase_ :int = os.path.join( os.path.dirname(lowercase__ ) , f"""bertarized_{os.path.basename(lowercase__ )}""" ) if not os.path.isdir(lowercase__ ): shutil.copytree(lowercase__ , lowercase__ ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(lowercase__ , os.path.join(lowercase__ , """pytorch_model.bin""" ) ) print("""\nPruned model saved! See you later!""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) __UpperCAmelCase = parser.parse_args() main(args)
84
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
0
'''simple docstring''' import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def UpperCamelCase_( snake_case : Optional[int] ): '''simple docstring''' snake_case_ = SwinConfig() snake_case_ = swin_name.split("_" ) snake_case_ = name_split[1] snake_case_ = int(name_split[4] ) snake_case_ = int(name_split[3][-1] ) if model_size == "tiny": snake_case_ = 9_6 snake_case_ = (2, 2, 6, 2) snake_case_ = (3, 6, 1_2, 2_4) elif model_size == "small": snake_case_ = 9_6 snake_case_ = (2, 2, 1_8, 2) snake_case_ = (3, 6, 1_2, 2_4) elif model_size == "base": snake_case_ = 1_2_8 snake_case_ = (2, 2, 1_8, 2) snake_case_ = (4, 8, 1_6, 3_2) else: snake_case_ = 1_9_2 snake_case_ = (2, 2, 1_8, 2) snake_case_ = (6, 1_2, 2_4, 4_8) if "in22k" in swin_name: snake_case_ = 2_1_8_4_1 else: snake_case_ = 1_0_0_0 snake_case_ = "huggingface/label-files" snake_case_ = "imagenet-1k-id2label.json" snake_case_ = json.load(open(hf_hub_download(snake_case , snake_case , repo_type="dataset" ) , "r" ) ) snake_case_ = {int(snake_case ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} snake_case_ = img_size snake_case_ = num_classes snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size return config def UpperCamelCase_( snake_case : Optional[int] ): '''simple docstring''' if "patch_embed.proj" in name: snake_case_ = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: snake_case_ = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: snake_case_ = "encoder." + name if "attn.proj" in name: snake_case_ = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: snake_case_ = name.replace("attn" , "attention.self" ) if "norm1" in name: snake_case_ = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: snake_case_ = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: snake_case_ = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: snake_case_ = name.replace("mlp.fc2" , "output.dense" ) if name == "norm.weight": snake_case_ = "layernorm.weight" if name == "norm.bias": snake_case_ = "layernorm.bias" if "head" in name: snake_case_ = name.replace("head" , "classifier" ) else: snake_case_ = "swin." + name return name def UpperCamelCase_( snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ = orig_state_dict.pop(snake_case ) if "mask" in key: continue elif "qkv" in key: snake_case_ = key.split("." ) snake_case_ = int(key_split[1] ) snake_case_ = int(key_split[3] ) snake_case_ = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[ dim : dim * 2, : ] snake_case_ = val[-dim:, :] else: snake_case_ = val[ :dim ] snake_case_ = val[ dim : dim * 2 ] snake_case_ = val[ -dim: ] else: snake_case_ = val return orig_state_dict def UpperCamelCase_( snake_case : str , snake_case : str ): '''simple docstring''' snake_case_ = timm.create_model(snake_case , pretrained=snake_case ) timm_model.eval() snake_case_ = get_swin_config(snake_case ) snake_case_ = SwinForImageClassification(snake_case ) model.eval() snake_case_ = convert_state_dict(timm_model.state_dict() , snake_case ) model.load_state_dict(snake_case ) snake_case_ = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ = AutoImageProcessor.from_pretrained("microsoft/{}".format(swin_name.replace("_" , "-" ) ) ) snake_case_ = Image.open(requests.get(snake_case , stream=snake_case ).raw ) snake_case_ = image_processor(images=snake_case , return_tensors="pt" ) snake_case_ = timm_model(inputs["pixel_values"] ) snake_case_ = model(**snake_case ).logits assert torch.allclose(snake_case , snake_case , atol=1e-3 ) print(f'Saving model {swin_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swin_name", default="swin_tiny_patch4_window7_224", type=str, help="Name of the Swin timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) _SCREAMING_SNAKE_CASE : Dict = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
85
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Tuple = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def __lowerCAmelCase (): print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
86
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = 1 _snake_case = 2 while i * i <= n: _snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _snake_case = 1 _snake_case = 1 while True: i += 1 t_num += i if count_divisors(__A ) > 500: break return t_num if __name__ == "__main__": print(solution())
42
0
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class snake_case_ ( unittest.TestCase ): def __UpperCamelCase ( self : List[str] ) -> int: lowercase__ : str = "ylacombe/bark-small" lowercase__ : Optional[int] = tempfile.mkdtemp() lowercase__ : str = "en_speaker_1" lowercase__ : List[str] = "This is a test string" lowercase__ : Dict = "speaker_embeddings_path.json" lowercase__ : List[str] = "speaker_embeddings" def __UpperCamelCase ( self : Optional[int] , **lowercase_ : Union[str, Any] ) -> Optional[int]: return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def __UpperCamelCase ( self : List[str] ) -> int: shutil.rmtree(self.tmpdirname ) def __UpperCamelCase ( self : List[str] ) -> List[Any]: lowercase__ : List[Any] = self.get_tokenizer() lowercase__ : Optional[int] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) lowercase__ : List[Any] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def __UpperCamelCase ( self : Tuple ) -> List[Any]: lowercase__ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) lowercase__ : str = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) lowercase__ : Any = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def __UpperCamelCase ( self : List[Any] ) -> List[str]: lowercase__ : int = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) lowercase__ : List[Any] = 35 lowercase__ : Any = 2 lowercase__ : List[str] = 8 lowercase__ : int = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset lowercase__ : List[str] = processor(text=self.input_string , voice_preset=lowercase_ ) lowercase__ : Union[str, Any] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file lowercase__ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) lowercase__ : Tuple = processor(text=self.input_string , voice_preset=lowercase_ ) lowercase__ : Dict = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub lowercase__ : Any = processor(text=self.input_string , voice_preset=self.voice_preset ) def __UpperCamelCase ( self : Optional[int] ) -> Any: lowercase__ : Tuple = self.get_tokenizer() lowercase__ : Optional[int] = BarkProcessor(tokenizer=lowercase_ ) lowercase__ : Dict = processor(text=self.input_string ) lowercase__ : int = tokenizer( self.input_string , padding="max_length" , max_length=2_56 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
87
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __UpperCAmelCase ( _lowerCamelCase ): @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _snake_case = self.get_env() _snake_case = '1' _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import AutoModel\n ' _snake_case = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
42
0
from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : List[str] , UpperCamelCase__ : int=13 , UpperCamelCase__ : Union[str, Any]=30 , UpperCamelCase__ : Optional[Any]=2 , UpperCamelCase__ : Optional[Any]=3 , UpperCamelCase__ : Any=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Tuple=32 , UpperCamelCase__ : Optional[Any]=2 , UpperCamelCase__ : Optional[int]=4 , UpperCamelCase__ : List[str]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Optional[int]=0.1 , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : int=10 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : Tuple=3 , UpperCamelCase__ : List[str]=0.6 , UpperCamelCase__ : Dict=None , ) -> str: """simple docstring""" __magic_name__ = parent __magic_name__ = batch_size __magic_name__ = image_size __magic_name__ = patch_size __magic_name__ = num_channels __magic_name__ = is_training __magic_name__ = use_labels __magic_name__ = hidden_size __magic_name__ = num_hidden_layers __magic_name__ = num_attention_heads __magic_name__ = intermediate_size __magic_name__ = hidden_act __magic_name__ = hidden_dropout_prob __magic_name__ = attention_probs_dropout_prob __magic_name__ = type_sequence_label_size __magic_name__ = initializer_range __magic_name__ = mask_ratio __magic_name__ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) __magic_name__ = (image_size // patch_size) ** 2 __magic_name__ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" __magic_name__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __magic_name__ = None if self.use_labels: __magic_name__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ = self.get_config() return config, pixel_values, labels def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def _lowercase ( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ) -> str: """simple docstring""" __magic_name__ = TFViTMAEModel(config=UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : Optional[int] ) -> Any: """simple docstring""" __magic_name__ = TFViTMAEForPreTraining(UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches __magic_name__ = (self.image_size // self.patch_size) ** 2 __magic_name__ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images __magic_name__ = 1 __magic_name__ = TFViTMAEForPreTraining(UpperCamelCase__ ) __magic_name__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __magic_name__ = model(UpperCamelCase__ , training=UpperCamelCase__ ) __magic_name__ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def _lowercase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __magic_name__ = self.prepare_config_and_inputs() ((__magic_name__) , (__magic_name__) , (__magic_name__)) = config_and_inputs __magic_name__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _A , _A , unittest.TestCase ): '''simple docstring''' a__ = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a__ = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a__ = False a__ = False a__ = False a__ = False def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" __magic_name__ = TFViTMAEModelTester(self ) __magic_name__ = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" pass def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) __magic_name__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) __magic_name__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __magic_name__ = [*signature.parameters.keys()] __magic_name__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def _lowercase ( self : List[Any] ) -> str: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _lowercase ( self : Any ) -> Optional[int]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" np.random.seed(2 ) __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = int((config.image_size // config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) __magic_name__ = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , noise=UpperCamelCase__ ) __magic_name__ = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) __magic_name__ = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) __magic_name__ = outputs_dict[0].numpy() __magic_name__ = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1E-6 ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" np.random.seed(2 ) __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = int((config.image_size // config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__ : int ): __magic_name__ = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): __magic_name__ = v.numpy() else: __magic_name__ = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) __magic_name__ = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = prepare_numpy_arrays(UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , noise=UpperCamelCase__ ) __magic_name__ = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self : str , UpperCamelCase__ : Tuple , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple ) -> List[Any]: """simple docstring""" np.random.seed(2 ) __magic_name__ = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __magic_name__ = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument __magic_name__ = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" np.random.seed(2 ) __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } __magic_name__ = int((config.image_size // config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __magic_name__ = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: __magic_name__ = main_layer_class(UpperCamelCase__ ) __magic_name__ = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } __magic_name__ = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) __magic_name__ = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) __magic_name__ = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) __magic_name__ = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" np.random.seed(2 ) __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = int((config.image_size // config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) __magic_name__ = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": __magic_name__ = outputs.last_hidden_state.numpy() __magic_name__ = 0 else: __magic_name__ = outputs.logits.numpy() __magic_name__ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) __magic_name__ = model_class.from_pretrained(UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": __magic_name__ = after_outputs["""last_hidden_state"""].numpy() __magic_name__ = 0 else: __magic_name__ = after_outputs["""logits"""].numpy() __magic_name__ = 0 __magic_name__ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1E-5 ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" np.random.seed(2 ) __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = int((config.image_size // config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __magic_name__ = model_class(UpperCamelCase__ ) __magic_name__ = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , noise=UpperCamelCase__ ) __magic_name__ = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) __magic_name__ = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config __magic_name__ = model_class.from_config(model.config ) __magic_name__ = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) __magic_name__ = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass @slow def _lowercase ( self : Any ) -> str: """simple docstring""" __magic_name__ = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def a__ ( ): '''simple docstring''' __magic_name__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def _lowercase ( self : Dict ) -> int: """simple docstring""" np.random.seed(2 ) __magic_name__ = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) __magic_name__ = self.default_image_processor __magic_name__ = prepare_img() __magic_name__ = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) __magic_name__ = ViTMAEConfig() __magic_name__ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) __magic_name__ = np.random.uniform(size=(1, num_patches) ) # forward pass __magic_name__ = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits __magic_name__ = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) __magic_name__ = tf.convert_to_tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1E-4 )
88
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCAmelCase_ ).to(lowerCAmelCase_ ) _snake_case = AutoTokenizer.from_pretrained('google/mt5-small' ) _snake_case = tokenizer('Hello there' , return_tensors='pt' ).input_ids _snake_case = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _snake_case = model(input_ids.to(lowerCAmelCase_ ) , labels=labels.to(lowerCAmelCase_ ) ).loss _snake_case = -(labels.shape[-1] * loss.item()) _snake_case = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
42
0
'''simple docstring''' import string def __lowerCamelCase ( lowerCAmelCase_ ) -> None: for key in range(len(string.ascii_uppercase ) ): _a : Union[str, Any] = '' for symbol in message: if symbol in string.ascii_uppercase: _a : Optional[Any] = string.ascii_uppercase.find(lowerCAmelCase_ ) _a : List[str] = num - key if num < 0: _a : str = num + len(string.ascii_uppercase ) _a : int = translated + string.ascii_uppercase[num] else: _a : Dict = translated + symbol print(f"""Decryption using Key #{key}: {translated}""" ) def __lowerCamelCase ( ) -> None: _a : int = input('Encrypted message: ' ) _a : Tuple = message.upper() decrypt(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
89
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : List[str] = { "configuration_pix2struct": [ "PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Pix2StructConfig", "Pix2StructTextConfig", "Pix2StructVisionConfig", ], "processing_pix2struct": ["Pix2StructProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = ["Pix2StructImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST", "Pix2StructPreTrainedModel", "Pix2StructForConditionalGeneration", "Pix2StructVisionModel", "Pix2StructTextModel", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowercase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case_ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> List[str]: '''simple docstring''' __lowerCamelCase = hf_hub_download( repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) __lowerCamelCase = VideoClassificationPipeline(model=lowerCamelCase__ , image_processor=lowerCamelCase__ , top_k=2 ) __lowerCamelCase = [ example_video_filepath, 'https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4', ] return video_classifier, examples def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ ) -> Tuple: '''simple docstring''' for example in examples: __lowerCamelCase = video_classifier(lowerCamelCase__ ) self.assertEqual( lowerCamelCase__ , [ {'score': ANY(lowerCamelCase__ ), 'label': ANY(lowerCamelCase__ )}, {'score': ANY(lowerCamelCase__ ), 'label': ANY(lowerCamelCase__ )}, ] , ) @require_torch def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = 'hf-internal-testing/tiny-random-VideoMAEForVideoClassification' __lowerCamelCase = VideoMAEFeatureExtractor( size={'shortest_edge': 10} , crop_size={'height': 10, 'width': 10} ) __lowerCamelCase = pipeline( 'video-classification' , model=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , frame_sampling_rate=4 ) __lowerCamelCase = hf_hub_download(repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) __lowerCamelCase = video_classifier(lowerCamelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=4 ) , [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}] , ) __lowerCamelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=4 ) , [ [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], ] , ) @require_tf def lowercase_ ( self ) -> Dict: '''simple docstring''' pass
90
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowercase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowercase : list[int] = [ord(letter) for letter in string.ascii_lowercase] lowercase : set[int] = {ord(char) for char in VALID_CHARS} lowercase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str | None: _snake_case = "" _snake_case = 42 _snake_case = 42 _snake_case = 42 for keychar, cipherchar in zip(cycle(__A ) , __A ): _snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__A ) return decoded def SCREAMING_SNAKE_CASE__ ( __A ) -> list[str]: _snake_case = [] for key in product(__A , repeat=3 ): _snake_case = try_key(__A , __A ) if encoded is not None: possibles.append(__A ) return possibles def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __A = "p059_cipher.txt" ) -> int: _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = 42 _snake_case = Path(__A ).parent.joinpath(__A ).read_text(encoding='utf-8' ) _snake_case = [int(__A ) for number in data.strip().split(',' )] _snake_case = filter_valid_chars(__A ) for common_word in COMMON_WORDS: _snake_case = filter_common_word(__A , __A ) if len(__A ) == 1: break _snake_case = possibles[0] return sum(ord(__A ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" from __future__ import annotations from random import choice def _A (__a ) -> Optional[int]: """simple docstring""" return choice(__a ) def _A (__a , __a ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = random_pivot(__a ) # partition based on pivot # linear time SCREAMING_SNAKE_CASE_ : Any = [e for e in lst if e < pivot] SCREAMING_SNAKE_CASE_ : Tuple = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
91
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): UpperCamelCase__ = True from torch.cuda.amp import autocast UpperCamelCase__ = logging.getLogger(__name__) def _a ( SCREAMING_SNAKE_CASE_ : List[Any]=None , SCREAMING_SNAKE_CASE_ : List[str]=None ): return field(default_factory=lambda: default , metadata=SCREAMING_SNAKE_CASE_ ) @dataclass class a__ : _a : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) _a : Optional[str] = field( default=snake_case__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _a : Optional[bool] = field( default=snake_case__ , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) _a : Optional[float] = field( default=0.1 , metadata={"""help""": """The dropout ratio for the attention probabilities."""} ) _a : Optional[float] = field( default=0.1 , metadata={"""help""": """The dropout ratio for activations inside the fully connected layer."""} ) _a : Optional[float] = field( default=0.1 , metadata={ """help""": """The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.""" } , ) _a : Optional[float] = field( default=0.1 , metadata={"""help""": """The dropout probabilitiy for all 1D convolutional layers in feature extractor."""} , ) _a : Optional[float] = field( default=0.05 , metadata={ """help""": ( """Propability of each feature vector along the time axis to be chosen as the start of the vector""" """span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature""" """vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.""" ) } , ) _a : Optional[float] = field(default=0.0 , metadata={"""help""": """The LayerDrop probability."""} ) @dataclass class a__ : _a : Optional[str] = field( default=snake_case__ , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) _a : Optional[str] = field( default="""train+validation""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) _a : bool = field( default=snake_case__ , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) _a : Optional[int] = field( default=snake_case__ , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) _a : Optional[int] = field( default=snake_case__ , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) _a : Optional[int] = field( default=snake_case__ , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of validation examples to this """ """value if set.""" ) } , ) _a : List[str] = list_field( default=[""",""", """?""", """.""", """!""", """-""", """;""", """:""", """\"\"""", """%""", """'""", """\"""", """�"""] , metadata={"""help""": """A list of characters to remove from the transcripts."""} , ) @dataclass class a__ : _a : WavaVecaProcessor _a : Union[bool, str] = True _a : Optional[int] = None _a : Optional[int] = None _a : Optional[int] = None _a : Optional[int] = None def __call__( self , _A ): """simple docstring""" __lowerCAmelCase = [{"input_values": feature["input_values"]} for feature in features] __lowerCAmelCase = [{"input_ids": feature["labels"]} for feature in features] __lowerCAmelCase = self.processor.pad( _A , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" , ) __lowerCAmelCase = self.processor.pad( labels=_A , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors="pt" , ) # replace padding with -100 to ignore loss correctly __lowerCAmelCase = labels_batch["input_ids"].masked_fill(labels_batch.attention_mask.ne(1 ) , -1_0_0 ) __lowerCAmelCase = labels return batch class a__ ( snake_case__ ): def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" model.train() __lowerCAmelCase = self._prepare_inputs(_A ) if self.use_amp: with autocast(): __lowerCAmelCase = self.compute_loss(_A , _A ) else: __lowerCAmelCase = self.compute_loss(_A , _A ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": __lowerCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": __lowerCAmelCase = loss.sum() / (inputs["labels"] >= 0).sum() else: raise ValueError(f"""{model.config.ctc_loss_reduction} is not valid. Choose one of ['mean', 'sum']""" ) if self.args.gradient_accumulation_steps > 1: __lowerCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(_A ).backward() elif self.use_apex: with amp.scale_loss(_A , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(_A ) else: loss.backward() return loss.detach() def _a ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __lowerCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = parser.parse_args_into_dataclasses() # Detecting last checkpoint. __lowerCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __lowerCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("Training/evaluation parameters %s" , SCREAMING_SNAKE_CASE_ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: __lowerCAmelCase = datasets.load_dataset( "common_voice" , data_args.dataset_config_name , split=data_args.train_split_name ) __lowerCAmelCase = datasets.load_dataset("common_voice" , data_args.dataset_config_name , split="test" ) # Create and save tokenizer __lowerCAmelCase = F"""[{''.join(data_args.chars_to_ignore )}]""" def remove_special_characters(SCREAMING_SNAKE_CASE_ : int ): __lowerCAmelCase = re.sub(SCREAMING_SNAKE_CASE_ , "" , batch["sentence"] ).lower() + " " return batch __lowerCAmelCase = train_dataset.map(SCREAMING_SNAKE_CASE_ , remove_columns=["sentence"] ) __lowerCAmelCase = eval_dataset.map(SCREAMING_SNAKE_CASE_ , remove_columns=["sentence"] ) def extract_all_chars(SCREAMING_SNAKE_CASE_ : Tuple ): __lowerCAmelCase = " ".join(batch["text"] ) __lowerCAmelCase = list(set(SCREAMING_SNAKE_CASE_ ) ) return {"vocab": [vocab], "all_text": [all_text]} __lowerCAmelCase = train_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , batch_size=-1 , keep_in_memory=SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , ) __lowerCAmelCase = train_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , batch_size=-1 , keep_in_memory=SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , ) __lowerCAmelCase = list(set(vocab_train["vocab"][0] ) | set(vocab_test["vocab"][0] ) ) __lowerCAmelCase = {v: k for k, v in enumerate(SCREAMING_SNAKE_CASE_ )} __lowerCAmelCase = vocab_dict[" "] del vocab_dict[" "] __lowerCAmelCase = len(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = len(SCREAMING_SNAKE_CASE_ ) with open("vocab.json" , "w" ) as vocab_file: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __lowerCAmelCase = WavaVecaCTCTokenizer( "vocab.json" , unk_token="[UNK]" , pad_token="[PAD]" , word_delimiter_token="|" , ) __lowerCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0.0 , do_normalize=SCREAMING_SNAKE_CASE_ , return_attention_mask=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction="mean" , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: __lowerCAmelCase = min(len(SCREAMING_SNAKE_CASE_ ) , data_args.max_train_samples ) __lowerCAmelCase = train_dataset.select(range(SCREAMING_SNAKE_CASE_ ) ) if data_args.max_val_samples is not None: __lowerCAmelCase = eval_dataset.select(range(data_args.max_val_samples ) ) __lowerCAmelCase = torchaudio.transforms.Resample(4_80_00 , 1_60_00 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(SCREAMING_SNAKE_CASE_ : str ): __lowerCAmelCase , __lowerCAmelCase = torchaudio.load(batch["path"] ) __lowerCAmelCase = resampler(SCREAMING_SNAKE_CASE_ ).squeeze().numpy() __lowerCAmelCase = 1_60_00 __lowerCAmelCase = batch["text"] return batch __lowerCAmelCase = train_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) __lowerCAmelCase = eval_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(SCREAMING_SNAKE_CASE_ : Tuple ): # check that all files have the correct sampling rate assert ( len(set(batch["sampling_rate"] ) ) == 1 ), F"""Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.""" __lowerCAmelCase = processor( audio=batch["speech"] , text=batch["target_text"] , sampling_rate=batch["sampling_rate"][0] ) batch.update(SCREAMING_SNAKE_CASE_ ) return batch __lowerCAmelCase = train_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , ) __lowerCAmelCase = eval_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , ) # Metric __lowerCAmelCase = datasets.load_metric("wer" ) def compute_metrics(SCREAMING_SNAKE_CASE_ : Tuple ): __lowerCAmelCase = pred.predictions __lowerCAmelCase = np.argmax(SCREAMING_SNAKE_CASE_ , axis=-1 ) __lowerCAmelCase = processor.tokenizer.pad_token_id __lowerCAmelCase = processor.batch_decode(SCREAMING_SNAKE_CASE_ ) # we do not want to group tokens when computing the metrics __lowerCAmelCase = processor.batch_decode(pred.label_ids , group_tokens=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = wer_metric.compute(predictions=SCREAMING_SNAKE_CASE_ , references=SCREAMING_SNAKE_CASE_ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator __lowerCAmelCase = DataCollatorCTCWithPadding(processor=SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ ) # Initialize our Trainer __lowerCAmelCase = CTCTrainer( model=SCREAMING_SNAKE_CASE_ , data_collator=SCREAMING_SNAKE_CASE_ , args=SCREAMING_SNAKE_CASE_ , compute_metrics=SCREAMING_SNAKE_CASE_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: __lowerCAmelCase = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): __lowerCAmelCase = model_args.model_name_or_path else: __lowerCAmelCase = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) __lowerCAmelCase = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE_ ) trainer.save_model() __lowerCAmelCase = train_result.metrics __lowerCAmelCase = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE_ ) ) __lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics("train" , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics("train" , SCREAMING_SNAKE_CASE_ ) trainer.save_state() # Evaluation __lowerCAmelCase = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) __lowerCAmelCase = trainer.evaluate() __lowerCAmelCase = data_args.max_val_samples if data_args.max_val_samples is not None else len(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics("eval" , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics("eval" , SCREAMING_SNAKE_CASE_ ) return results if __name__ == "__main__": main()
92
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
'''simple docstring''' def snake_case_ ( __SCREAMING_SNAKE_CASE : int = 10 , __SCREAMING_SNAKE_CASE : int = 22 ): """simple docstring""" lowercase_ : Tuple = range(1 , __SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = range(1 , __SCREAMING_SNAKE_CASE ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(f"""{solution(1_0, 2_2) = }""")
93
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from jiwer import compute_measures import datasets snake_case : List[Any] = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' snake_case : int = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' snake_case : str = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=False ): if concatenate_texts: return compute_measures(_lowerCamelCase , _lowerCamelCase )["wer"] else: a :Tuple = 0 a :str = 0 for prediction, reference in zip(_lowerCamelCase , _lowerCamelCase ): a :Tuple = compute_measures(_lowerCamelCase , _lowerCamelCase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
94
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize("dataset_size" , [None, 400 * 2**20, 600 * 2**20] ) @pytest.mark.parametrize("input_in_memory_max_size" , ["default", 0, 100 * 2**20, 900 * 2**20] ) def _A ( SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , "IN_MEMORY_MAX_SIZE" , SCREAMING_SNAKE_CASE ) a__ : Union[str, Any] =datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: a__ : List[str] =dataset_size < in_memory_max_size else: a__ : Any =False a__ : List[str] =is_small_dataset(SCREAMING_SNAKE_CASE ) assert result == expected
95
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = CycleDiffusionPipeline lowerCamelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """negative_prompt""", """height""", """width""", """negative_prompt_embeds""", } lowerCamelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCamelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""} ) lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) _lowerCamelCase : Optional[Any] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , num_train_timesteps=1000 , clip_sample=lowercase , set_alpha_to_one=lowercase , ) torch.manual_seed(0 ) _lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCamelCase : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) _lowerCamelCase : Any = CLIPTextModel(lowercase ) _lowerCamelCase : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCamelCase : int = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self , lowercase , lowercase=0 ): _lowerCamelCase : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) _lowerCamelCase : Optional[Any] = image / 2 + 0.5 if str(lowercase ).startswith('mps' ): _lowerCamelCase : Optional[int] = torch.manual_seed(lowercase ) else: _lowerCamelCase : Union[str, Any] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : List[Any] = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase : str = self.get_dummy_components() _lowerCamelCase : Tuple = CycleDiffusionPipeline(**lowercase ) _lowerCamelCase : int = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[Any] = self.get_dummy_inputs(lowercase ) _lowerCamelCase : List[Any] = pipe(**lowercase ) _lowerCamelCase : Optional[int] = output.images _lowerCamelCase : Optional[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) _lowerCamelCase : List[str] = np.array([0.44_59, 0.49_43, 0.45_44, 0.66_43, 0.54_74, 0.43_27, 0.57_01, 0.59_59, 0.51_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.get_dummy_components() for name, module in components.items(): if hasattr(lowercase , 'half' ): _lowerCamelCase : str = module.half() _lowerCamelCase : List[str] = CycleDiffusionPipeline(**lowercase ) _lowerCamelCase : Union[str, Any] = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[Any] = self.get_dummy_inputs(lowercase ) _lowerCamelCase : Union[str, Any] = pipe(**lowercase ) _lowerCamelCase : Tuple = output.images _lowerCamelCase : int = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) _lowerCamelCase : str = np.array([0.35_06, 0.45_43, 0.4_46, 0.45_75, 0.51_95, 0.41_55, 0.52_73, 0.5_18, 0.41_16] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def A_ ( self ): return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def A_ ( self ): return super().test_inference_batch_single_identical() @skip_mps def A_ ( self ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def A_ ( self ): return super().test_save_load_optional_components() @skip_mps def A_ ( self ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowerCamelCase : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) _lowerCamelCase : Optional[Any] = init_image.resize((512, 512) ) _lowerCamelCase : Dict = 'CompVis/stable-diffusion-v1-4' _lowerCamelCase : Any = DDIMScheduler.from_pretrained(lowercase , subfolder='scheduler' ) _lowerCamelCase : Optional[int] = CycleDiffusionPipeline.from_pretrained( lowercase , scheduler=lowercase , safety_checker=lowercase , torch_dtype=torch.floataa , revision='fp16' ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : Tuple = 'A black colored car' _lowerCamelCase : Optional[Any] = 'A blue colored car' _lowerCamelCase : List[Any] = torch.manual_seed(0 ) _lowerCamelCase : str = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type='np' , ) _lowerCamelCase : Optional[Any] = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def A_ ( self ): _lowerCamelCase : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowerCamelCase : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) _lowerCamelCase : Tuple = init_image.resize((512, 512) ) _lowerCamelCase : List[Any] = 'CompVis/stable-diffusion-v1-4' _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase , subfolder='scheduler' ) _lowerCamelCase : str = CycleDiffusionPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : List[str] = 'A black colored car' _lowerCamelCase : Tuple = 'A blue colored car' _lowerCamelCase : Dict = torch.manual_seed(0 ) _lowerCamelCase : Dict = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type='np' , ) _lowerCamelCase : str = output.images assert np.abs(image - expected_image ).max() < 2E-2
96
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { '''CarlCochet/trajectory-transformer-halfcheetah-medium-v2''': ( '''https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json''' ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class lowercase ( A__ ): """simple docstring""" _a = 'trajectory_transformer' _a = ['past_key_values'] _a = { 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , UpperCamelCase_=100 , UpperCamelCase_=5 , UpperCamelCase_=1 , UpperCamelCase_=1 , UpperCamelCase_=249 , UpperCamelCase_=6 , UpperCamelCase_=17 , UpperCamelCase_=25 , UpperCamelCase_=4 , UpperCamelCase_=4 , UpperCamelCase_=128 , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=0.0006 , UpperCamelCase_=512 , UpperCamelCase_=0.02 , UpperCamelCase_=1e-12 , UpperCamelCase_=1 , UpperCamelCase_=True , UpperCamelCase_=1 , UpperCamelCase_=50256 , UpperCamelCase_=50256 , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :int = vocab_size UpperCamelCase__ :str = action_weight UpperCamelCase__ :Dict = reward_weight UpperCamelCase__ :Optional[int] = value_weight UpperCamelCase__ :List[Any] = max_position_embeddings UpperCamelCase__ :int = block_size UpperCamelCase__ :Optional[Any] = action_dim UpperCamelCase__ :Union[str, Any] = observation_dim UpperCamelCase__ :int = transition_dim UpperCamelCase__ :int = learning_rate UpperCamelCase__ :int = n_layer UpperCamelCase__ :int = n_head UpperCamelCase__ :List[Any] = n_embd UpperCamelCase__ :Optional[int] = embd_pdrop UpperCamelCase__ :Dict = attn_pdrop UpperCamelCase__ :List[str] = resid_pdrop UpperCamelCase__ :Optional[Any] = initializer_range UpperCamelCase__ :Any = layer_norm_eps UpperCamelCase__ :Optional[int] = kaiming_initializer_range UpperCamelCase__ :Dict = use_cache super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ )
97
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
"""simple docstring""" import os import unittest from transformers.models.phobert.tokenization_phobert import VOCAB_FILES_NAMES, PhobertTokenizer from ...test_tokenization_common import TokenizerTesterMixin class snake_case ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" snake_case__ = PhobertTokenizer snake_case__ = False def __lowerCAmelCase ( self : str ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ = ['T@@', 'i', 'I', 'R@@', 'r', 'e@@'] UpperCAmelCase__ = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) UpperCAmelCase__ = ['#version: 0.2', 'l à</w>'] UpperCAmelCase__ = {'unk_token': '<unk>'} UpperCAmelCase__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['vocab_file'] ) UpperCAmelCase__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file ,'w' ,encoding='utf-8' ) as fp: for token in vocab_tokens: fp.write(f'''{token} {vocab_tokens[token]}\n''' ) with open(self.merges_file ,'w' ,encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowerCAmelCase ( self : List[Any] ,**lowerCamelCase__ : List[Any] ): kwargs.update(self.special_tokens_map ) return PhobertTokenizer.from_pretrained(self.tmpdirname ,**lowerCamelCase__ ) def __lowerCAmelCase ( self : int ,lowerCamelCase__ : Union[str, Any] ): UpperCAmelCase__ = 'Tôi là VinAI Research' UpperCAmelCase__ = 'T<unk> i <unk> <unk> <unk> <unk> <unk> <unk> I Re<unk> e<unk> <unk> <unk> <unk>' return input_text, output_text def __lowerCAmelCase ( self : Optional[int] ): UpperCAmelCase__ = PhobertTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) UpperCAmelCase__ = 'Tôi là VinAI Research' UpperCAmelCase__ = 'T@@ ô@@ i l@@ à V@@ i@@ n@@ A@@ I R@@ e@@ s@@ e@@ a@@ r@@ c@@ h'.split() UpperCAmelCase__ = tokenizer.tokenize(lowerCamelCase__ ) print(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) UpperCAmelCase__ = tokens + [tokenizer.unk_token] UpperCAmelCase__ = [4, 3, 5, 3, 3, 3, 3, 3, 3, 6, 7, 9, 3, 9, 3, 3, 3, 3, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,lowerCamelCase__ )
98
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class A__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" __A : Any = StableUnCLIPImgaImgPipeline __A : Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS __A : Dict = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __A : Dict = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __A : List[Any] = frozenset([] ) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : Any = 32 a__ : Any = embedder_hidden_size # image encoding components a__ : Optional[Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) a__ : Tuple = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=lowercase , projection_dim=lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) a__ : Union[str, Any] = StableUnCLIPImageNormalizer(embedding_dim=lowercase) a__ : str = DDPMScheduler(beta_schedule='squaredcos_cap_v2') torch.manual_seed(0) a__ : Tuple = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') torch.manual_seed(0) a__ : List[str] = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , )) torch.manual_seed(0) a__ : Optional[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'CrossAttnUpBlock2D') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='projection' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=lowercase , layers_per_block=1 , upcast_attention=lowercase , use_linear_projection=lowercase , ) torch.manual_seed(0) a__ : List[str] = DDIMScheduler( beta_schedule='scaled_linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='v_prediction' , set_alpha_to_one=lowercase , steps_offset=1 , ) torch.manual_seed(0) a__ : str = AutoencoderKL() a__ : int = { # image encoding components 'feature_extractor': feature_extractor, 'image_encoder': image_encoder.eval(), # image noising components 'image_normalizer': image_normalizer.eval(), 'image_noising_scheduler': image_noising_scheduler, # regular denoising components 'tokenizer': tokenizer, 'text_encoder': text_encoder.eval(), 'unet': unet.eval(), 'scheduler': scheduler, 'vae': vae.eval(), } return components def __lowercase ( self , lowercase , lowercase=0 , lowercase=True) -> Union[str, Any]: '''simple docstring''' if str(lowercase).startswith('mps'): a__ : List[Any] = torch.manual_seed(lowercase) else: a__ : Optional[Any] = torch.Generator(device=lowercase).manual_seed(lowercase) a__ : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase)).to(lowercase) if pil_image: a__ : Any = input_image * 0.5 + 0.5 a__ : int = input_image.clamp(0 , 1) a__ : Any = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() a__ : Tuple = DiffusionPipeline.numpy_to_pil(lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def __lowercase ( self) -> Dict: '''simple docstring''' a__ : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator a__ : str = self.get_dummy_components() a__ : Optional[Any] = StableUnCLIPImgaImgPipeline(**lowercase) a__ : Union[str, Any] = sd_pipe.to(lowercase) sd_pipe.set_progress_bar_config(disable=lowercase) a__ : str = self.get_dummy_inputs(lowercase) inputs.update({'image_embeds': None}) a__ : int = sd_pipe(**lowercase).images a__ : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) a__ : List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : str = torch_device in ['cpu', 'mps'] self._test_attention_slicing_forward_pass(test_max_difference=lowercase) def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : int = torch_device in ['cpu', 'mps'] self._test_inference_batch_single_identical(test_max_difference=lowercase) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __lowercase ( self) -> int: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_max_difference=lowercase) @slow @require_torch_gpu class A__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png') a__ : int = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy') a__ : Any = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-l-img2img' , torch_dtype=torch.floataa) pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a__ : Any = torch.Generator(device='cpu').manual_seed(0) a__ : Dict = pipe(lowercase , 'anime turle' , generator=lowercase , output_type='np') a__ : Any = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowercase , lowercase) def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png') a__ : Tuple = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy') a__ : int = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-h-img2img' , torch_dtype=torch.floataa) pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a__ : Dict = torch.Generator(device='cpu').manual_seed(0) a__ : Dict = pipe(lowercase , 'anime turle' , generator=lowercase , output_type='np') a__ : Any = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowercase , lowercase) def __lowercase ( self) -> List[str]: '''simple docstring''' a__ : Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() a__ : Optional[int] = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-h-img2img' , torch_dtype=torch.floataa) a__ : Optional[int] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a__ : List[Any] = pipe( lowercase , 'anime turtle' , num_inference_steps=2 , output_type='np' , ) a__ : Union[str, Any] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
99
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): # Check if the input is valid if not len(UpperCamelCase_ ) == len(UpperCamelCase_ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = equationa __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = equationa # Calculate the determinants of the matrices __SCREAMING_SNAKE_CASE = aa * ba - aa * ba __SCREAMING_SNAKE_CASE = ca * ba - ca * ba __SCREAMING_SNAKE_CASE = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: __SCREAMING_SNAKE_CASE = determinant_x / determinant __SCREAMING_SNAKE_CASE = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
100
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0