code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
__lowerCamelCase : Dict = {} def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[str] , __UpperCamelCase : Optional[int] , __UpperCamelCase : Optional[Any] ) -> int: """simple docstring""" if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on SCREAMING_SNAKE_CASE__ = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one SCREAMING_SNAKE_CASE__ = _calculate(days - 1 , __lowerCAmelCase , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 SCREAMING_SNAKE_CASE__ = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter SCREAMING_SNAKE_CASE__ = _calculate(days - 1 , __lowerCAmelCase , 0 ) SCREAMING_SNAKE_CASE__ = state_late + state_absent + state_ontime SCREAMING_SNAKE_CASE__ = prizestrings return prizestrings def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Union[str, Any] = 30 ) -> int: """simple docstring""" return _calculate(__lowerCAmelCase , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
219
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _a = get_logger(__name__) class __lowerCamelCase ( enum.Enum): """simple docstring""" UpperCamelCase__ = "all_checks" UpperCamelCase__ = "basic_checks" UpperCamelCase__ = "no_checks" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None )-> str: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase = ' for ' + verification_name if verification_name is not None else '' if len(__lowerCAmelCase ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__lowerCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__lowerCAmelCase ) ) logger.info('All the splits matched successfully.' ) def __A ( __lowerCAmelCase , __lowerCAmelCase = True )-> dict: """simple docstring""" if record_checksum: _UpperCAmelCase = shaaaa() with open(__lowerCAmelCase , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__lowerCAmelCase ) _UpperCAmelCase = m.hexdigest() else: _UpperCAmelCase = None return {"num_bytes": os.path.getsize(__lowerCAmelCase ), "checksum": checksum} def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
39
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { """shi-labs/nat-mini-in1k-224""": """https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json""", # See all Nat models at https://huggingface.co/models?filter=nat } class a__ ( snake_case__ , snake_case__ ): """simple docstring""" __lowerCamelCase = 'nat' __lowerCamelCase = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , lowercase=4 , lowercase=3 , lowercase=64 , lowercase=[3, 4, 6, 5] , lowercase=[2, 4, 8, 16] , lowercase=7 , lowercase=3.0 , lowercase=True , lowercase=0.0 , lowercase=0.0 , lowercase=0.1 , lowercase="gelu" , lowercase=0.02 , lowercase=1e-5 , lowercase=0.0 , lowercase=None , lowercase=None , **lowercase , ) -> int: '''simple docstring''' super().__init__(**lowercase ) A__ = patch_size A__ = num_channels A__ = embed_dim A__ = depths A__ = len(lowercase ) A__ = num_heads A__ = kernel_size A__ = mlp_ratio A__ = qkv_bias A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = drop_path_rate A__ = hidden_act A__ = layer_norm_eps A__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model A__ = int(embed_dim * 2 ** (len(lowercase ) - 1) ) A__ = layer_scale_init_value A__ = ["stem"] + [F'stage{idx}' for idx in range(1 , len(lowercase ) + 1 )] A__ , A__ = get_aligned_output_features_output_indices( out_features=lowercase , out_indices=lowercase , stage_names=self.stage_names )
68
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=16 , UpperCAmelCase=[1, 2, 1] , UpperCAmelCase=[2, 2, 4] , UpperCAmelCase=2 , UpperCAmelCase=2.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=10 , UpperCAmelCase=8 , UpperCAmelCase=["stage1", "stage2", "stage3"] , UpperCAmelCase=[1, 2, 3] , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = embed_dim _UpperCAmelCase = depths _UpperCAmelCase = num_heads _UpperCAmelCase = window_size _UpperCAmelCase = mlp_ratio _UpperCAmelCase = qkv_bias _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = drop_path_rate _UpperCAmelCase = hidden_act _UpperCAmelCase = use_absolute_embeddings _UpperCAmelCase = patch_norm _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = initializer_range _UpperCAmelCase = is_training _UpperCAmelCase = scope _UpperCAmelCase = use_labels _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = encoder_stride _UpperCAmelCase = out_features _UpperCAmelCase = out_indices def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) _UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCAmelCase ): _UpperCAmelCase = ['stem'] _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCamelCase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) # Swin has a different seq_length _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = 3 _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCAmelCase ): _UpperCAmelCase = 0 return t def check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase={} ): with torch.no_grad(): _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ).to_tuple() def recursive_check(UpperCAmelCase , UpperCAmelCase ): if isinstance(UpperCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase , UpperCAmelCase ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCAmelCase ) , set_nan_tensor_to_zero(UpperCAmelCase ) , atol=1e-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}. Dict has""" F""" `nan`: {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}.""" ) , ) recursive_check(UpperCAmelCase , UpperCAmelCase ) for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) @require_torch class __lowerCamelCase ( unittest.TestCase , snake_case__): """simple docstring""" UpperCamelCase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCamelCase__ = MaskFormerSwinConfig def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: _UpperCAmelCase = backbone_class(UpperCAmelCase ) backbone.to(UpperCAmelCase ) backbone.eval() _UpperCAmelCase = backbone(**UpperCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _UpperCAmelCase = backbone(**UpperCAmelCase , output_hidden_states=UpperCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _UpperCAmelCase = backbone(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertIsNotNone(outputs.attentions )
39
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCAmelCase ( snake_case__ , snake_case__ , snake_case__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : List[Any] = StableDiffusionInpaintPipeline __UpperCamelCase : int = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __UpperCamelCase : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCamelCase : str = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __UpperCamelCase : Optional[Any] = frozenset([] ) def __magic_name__ ( self : Any ): """simple docstring""" torch.manual_seed(0 ) _A: Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=lowerCAmelCase_ , ) _A: Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) _A: Optional[Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _A: List[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) _A: Any = CLIPTextModel(lowerCAmelCase_ ) _A: Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _A: Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def __magic_name__ ( self : str , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple=0 ): """simple docstring""" _A: Union[str, Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) _A: int = image.cpu().permute(0 , 2 , 3 , 1 )[0] _A: Dict = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) _A: Any = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((6_4, 6_4) ) if str(lowerCAmelCase_ ).startswith('''mps''' ): _A: Optional[int] = torch.manual_seed(lowerCAmelCase_ ) else: _A: int = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _A: Tuple = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def __magic_name__ ( self : List[str] ): """simple docstring""" _A: List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _A: Optional[Any] = self.get_dummy_components() _A: Any = StableDiffusionInpaintPipeline(**lowerCAmelCase_ ) _A: List[Any] = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A: str = self.get_dummy_inputs(lowerCAmelCase_ ) _A: str = sd_pipe(**lowerCAmelCase_ ).images _A: Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _A: Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __magic_name__ ( self : Optional[int] ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __magic_name__ ( self : Optional[Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self : str ): """simple docstring""" _A: Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _A: Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _A: Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) _A: Tuple = '''stabilityai/stable-diffusion-2-inpainting''' _A: Optional[int] = StableDiffusionInpaintPipeline.from_pretrained(lowerCAmelCase_ , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() _A: int = '''Face of a yellow cat, high resolution, sitting on a park bench''' _A: Optional[int] = torch.manual_seed(0 ) _A: int = pipe( prompt=lowerCAmelCase_ , image=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , generator=lowerCAmelCase_ , output_type='''np''' , ) _A: Optional[Any] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9e-3 def __magic_name__ ( self : str ): """simple docstring""" _A: Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _A: List[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _A: Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) _A: Tuple = '''stabilityai/stable-diffusion-2-inpainting''' _A: Optional[int] = StableDiffusionInpaintPipeline.from_pretrained( lowerCAmelCase_ , torch_dtype=torch.floataa , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() _A: Optional[Any] = '''Face of a yellow cat, high resolution, sitting on a park bench''' _A: int = torch.manual_seed(0 ) _A: Union[str, Any] = pipe( prompt=lowerCAmelCase_ , image=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , generator=lowerCAmelCase_ , output_type='''np''' , ) _A: Dict = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5e-1 def __magic_name__ ( self : Optional[Any] ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _A: List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) _A: int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) _A: Dict = '''stabilityai/stable-diffusion-2-inpainting''' _A: str = PNDMScheduler.from_pretrained(lowerCAmelCase_ , subfolder='''scheduler''' ) _A: Tuple = StableDiffusionInpaintPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , scheduler=lowerCAmelCase_ , torch_dtype=torch.floataa , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _A: Dict = '''Face of a yellow cat, high resolution, sitting on a park bench''' _A: str = torch.manual_seed(0 ) _A: str = pipe( prompt=lowerCAmelCase_ , image=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , generator=lowerCAmelCase_ , num_inference_steps=2 , output_type='''np''' , ) _A: Any = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
121
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = TransfoXLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" super().setUp() _UpperCAmelCase = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = '<unk> UNwanted , running' _UpperCAmelCase = '<unk> unwanted, running' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(UpperCAmelCase , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [0, 4, 8, 7] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) _UpperCAmelCase = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' _UpperCAmelCase = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(UpperCAmelCase ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(UpperCAmelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
39
0
"""simple docstring""" class _UpperCAmelCase: def __init__( self , __a) -> Tuple: '''simple docstring''' _UpperCamelCase = val _UpperCamelCase = None _UpperCamelCase = None def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' if self.val: if val < self.val: if self.left is None: _UpperCamelCase = Node(__a) else: self.left.insert(__a) elif val > self.val: if self.right is None: _UpperCamelCase = Node(__a) else: self.right.insert(__a) else: _UpperCamelCase = val def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[Any]: """simple docstring""" if root: inorder(root.left, __lowerCAmelCase ) res.append(root.val ) inorder(root.right, __lowerCAmelCase ) def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" if len(__lowerCAmelCase ) == 0: return arr _UpperCamelCase = Node(arr[0] ) for i in range(1, len(__lowerCAmelCase ) ): root.insert(arr[i] ) # Traverse BST in order. _UpperCamelCase = [] inorder(__lowerCAmelCase, __lowerCAmelCase ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
194
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = {'''configuration_opt''': ['''OPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OPTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''OPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OPTForCausalLM''', '''OPTModel''', '''OPTPreTrainedModel''', '''OPTForSequenceClassification''', '''OPTForQuestionAnswering''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['''TFOPTForCausalLM''', '''TFOPTModel''', '''TFOPTPreTrainedModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''FlaxOPTForCausalLM''', '''FlaxOPTModel''', '''FlaxOPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
39
0
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def lowerCamelCase_ ( ) -> Tuple: """simple docstring""" raise RuntimeError('''CUDA out of memory.''' ) class __lowerCAmelCase ( nn.Module ): def __init__(self ) -> Tuple: '''simple docstring''' super().__init__() snake_case_ : str = nn.Linear(3 , 4 ) snake_case_ : Union[str, Any] = nn.BatchNormad(4 ) snake_case_ : Optional[Any] = nn.Linear(4 , 5 ) def lowerCamelCase (self , __magic_name__ ) -> str: '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(__magic_name__ ) ) ) class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ ): nonlocal batch_sizes batch_sizes.append(__magic_name__ ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(__magic_name__ , [128, 64, 32, 16, 8] ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : Any = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ , __magic_name__ ): nonlocal batch_sizes batch_sizes.append(__magic_name__ ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga snake_case_ , snake_case_ : int = mock_training_loop_function('''hello''' ) self.assertListEqual(__magic_name__ , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, '''hello'''] ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(__magic_name__ ): pass with self.assertRaises(__magic_name__ ) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0] ) def lowerCamelCase (self ) -> Any: '''simple docstring''' @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(__magic_name__ ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(__magic_name__ ) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0] ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ , __magic_name__ , __magic_name__ ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(__magic_name__ ) as cm: mock_training_loop_function(128 , '''hello''' , '''world''' ) self.assertIn('''Batch size was passed into `f`''' , cm.exception.args[0] ) self.assertIn('''`f(arg1=\'hello\', arg2=\'world\')''' , cm.exception.args[0] ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(__magic_name__ ): raise ValueError('''Oops, we had an error!''' ) with self.assertRaises(__magic_name__ ) as cm: mock_training_loop_function() self.assertIn('''Oops, we had an error!''' , cm.exception.args[0] ) @require_cuda def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = torch.cuda.memory_allocated() snake_case_ : Any = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , __magic_name__ ) snake_case_ : Optional[Any] = release_memory(__magic_name__ ) self.assertEqual(torch.cuda.memory_allocated() , __magic_name__ )
279
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" _UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ), F"""{len(__lowerCAmelCase )} != {len(__lowerCAmelCase )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) _a = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _a = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def __A ( __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" try: _UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(__lowerCAmelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __A ( __lowerCAmelCase , __lowerCAmelCase = "student" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , )-> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" _UpperCAmelCase = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowerCAmelCase , __lowerCAmelCase ): AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) # purely for convenience _UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(__lowerCAmelCase ).eval() else: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), F"""teacher must be a model or string got type {type(__lowerCAmelCase )}""" _UpperCAmelCase = teacher.config.to_diff_dict() try: _UpperCAmelCase , _UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowerCAmelCase ) # Copy weights _UpperCAmelCase = teacher.config_class(**__lowerCAmelCase ) _UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. _UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=__lowerCAmelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save _UpperCAmelCase , _UpperCAmelCase = list(range(__lowerCAmelCase ) ), list(range(__lowerCAmelCase ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(__lowerCAmelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) if d_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) try: if hasattr( __lowerCAmelCase , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowerCAmelCase ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowerCAmelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowerCAmelCase ) copy_layers(teacher.decoder.block , student.decoder.block , __lowerCAmelCase ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) _UpperCAmelCase = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowerCAmelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
39
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class UpperCamelCase__ ( snake_case__ ): _SCREAMING_SNAKE_CASE : Optional[Any] = 42 class UpperCamelCase__ ( snake_case__ ,snake_case__ ): @register_to_config def __init__(self : int , snake_case_ : List[str] = 3 , snake_case_ : Dict = 3 , snake_case_ : List[str] = ("DownEncoderBlock2D",) , snake_case_ : Optional[Any] = ("UpDecoderBlock2D",) , snake_case_ : str = (6_4,) , snake_case_ : Optional[Any] = 1 , snake_case_ : int = "silu" , snake_case_ : Union[str, Any] = 3 , snake_case_ : str = 3_2 , snake_case_ : Union[str, Any] = 2_5_6 , snake_case_ : Dict = 3_2 , snake_case_ : Tuple = None , snake_case_ : List[Any] = 0.1_8215 , snake_case_ : Any = "group" , ): super().__init__() # pass init params to Encoder __a : List[str] = Encoder( in_channels=snake_case_ , out_channels=snake_case_ , down_block_types=snake_case_ , block_out_channels=snake_case_ , layers_per_block=snake_case_ , act_fn=snake_case_ , norm_num_groups=snake_case_ , double_z=snake_case_ , ) __a : List[Any] = vq_embed_dim if vq_embed_dim is not None else latent_channels __a : Any = nn.Convad(snake_case_ , snake_case_ , 1 ) __a : int = VectorQuantizer(snake_case_ , snake_case_ , beta=0.25 , remap=snake_case_ , sane_index_shape=snake_case_ ) __a : int = nn.Convad(snake_case_ , snake_case_ , 1 ) # pass init params to Decoder __a : Optional[Any] = Decoder( in_channels=snake_case_ , out_channels=snake_case_ , up_block_types=snake_case_ , block_out_channels=snake_case_ , layers_per_block=snake_case_ , act_fn=snake_case_ , norm_num_groups=snake_case_ , norm_type=snake_case_ , ) @apply_forward_hook def lowerCAmelCase (self : str , snake_case_ : Dict , snake_case_ : Optional[Any] = True ): __a : Optional[Any] = self.encoder(snake_case_ ) __a : Union[str, Any] = self.quant_conv(snake_case_ ) if not return_dict: return (h,) return VQEncoderOutput(latents=snake_case_ ) @apply_forward_hook def lowerCAmelCase (self : Optional[Any] , snake_case_ : Tuple , snake_case_ : Optional[Any] = False , snake_case_ : str = True ): if not force_not_quantize: __a , __a , __a : List[Any] = self.quantize(snake_case_ ) else: __a : Any = h __a : int = self.post_quant_conv(snake_case_ ) __a : Tuple = self.decoder(snake_case_ , quant if self.config.norm_type == '''spatial''' else None ) if not return_dict: return (dec,) return DecoderOutput(sample=snake_case_ ) def lowerCAmelCase (self : Dict , snake_case_ : int , snake_case_ : Optional[Any] = True ): __a : List[Any] = sample __a : List[str] = self.encode(snake_case_ ).latents __a : List[str] = self.decode(snake_case_ ).sample if not return_dict: return (dec,) return DecoderOutput(sample=snake_case_ )
216
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False )-> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase = '' else: _UpperCAmelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase = in_proj_bias[: config.hidden_size] _UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase = in_proj_bias[-config.hidden_size :] def __A ( __lowerCAmelCase )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" _UpperCAmelCase = dct.pop(__lowerCAmelCase ) _UpperCAmelCase = val def __A ( )-> str: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True )-> List[str]: """simple docstring""" _UpperCAmelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCAmelCase = 8 # set labels if required if not base_model: _UpperCAmelCase = 1_000 _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCAmelCase = 384 _UpperCAmelCase = 1_536 _UpperCAmelCase = 12 _UpperCAmelCase = 6 # load original model from torch hub _UpperCAmelCase = torch.hub.load('facebookresearch/dino:main' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCAmelCase = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _UpperCAmelCase = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _UpperCAmelCase = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCAmelCase = ViTImageProcessor() _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCAmelCase = encoding['pixel_values'] _UpperCAmelCase = model(__lowerCAmelCase ) if base_model: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) _a = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
39
0
from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
def _UpperCamelCase ( snake_case__ ) -> bool: if not all(x.isalpha() for x in string ): raise ValueError("String must only contain alphabetic characters." ) __UpperCAmelCase : Tuple = sorted(string.lower() ) return len(__lowerCAmelCase ) == len(set(__lowerCAmelCase ) ) if __name__ == "__main__": _snake_case = input('''Enter a string ''').strip() _snake_case = is_isogram(input_str) print(F'{input_str} is {"an" if isogram else "not an"} isogram.')
157
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=3 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=10 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[1, 1, 2, 1] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase="relu" , UpperCAmelCase=3 , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = embeddings_size _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = scope _UpperCAmelCase = len(UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = TFResNetModel(config=UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFResNetForImageClassification(UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase = layer_type _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) @slow def UpperCamelCase ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TFResNetModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __A ( )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" @cached_property def UpperCamelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=UpperCAmelCase , return_tensors='tf' ) # forward pass _UpperCAmelCase = model(**UpperCAmelCase ) # verify the logits _UpperCAmelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _UpperCAmelCase = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCAmelCase , atol=1e-4 ) )
39
0
'''simple docstring''' import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def __a(): '''simple docstring''' with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(__lowerCAmelCase ): requests.request("GET" , "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET" , "https://huggingface.co" , timeout=1.0 ) @pytest.mark.integration def __a(): '''simple docstring''' with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET" , "https://huggingface.co" ) def __a(): '''simple docstring''' with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(__lowerCAmelCase ): http_head("https://huggingface.co" )
158
def __A ( __lowerCAmelCase )-> list: """simple docstring""" if len(__lowerCAmelCase ) < 2: return collection def circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , mid + 1 , __lowerCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) - 1 ) return collection if __name__ == "__main__": _a = input('''Enter numbers separated by a comma:\n''').strip() _a = [int(item) for item in user_input.split(''',''')] print(circle_sort(unsorted))
39
0
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = inspect.getfile(accelerate.test_utils ) lowerCAmelCase__ : List[str] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['scripts', 'external_deps', 'test_metrics.py'] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 lowerCAmelCase__ : str = test_metrics @require_cpu def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' debug_launcher(self.test_metrics.main ) @require_single_gpu def _lowerCamelCase ( self : str ): '''simple docstring''' self.test_metrics.main() @require_multi_gpu def _lowerCamelCase ( self : List[str] ): '''simple docstring''' print(f'''Found {torch.cuda.device_count()} devices.''' ) lowerCAmelCase__ : List[str] = ['torchrun', f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a , env=os.environ.copy() )
212
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "Pix2StructImageProcessor" UpperCamelCase__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = False super().__init__(UpperCAmelCase , UpperCAmelCase ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 2048 , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , **UpperCAmelCase ) else: # add pixel_values and bbox _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , header_text=UpperCAmelCase , **UpperCAmelCase ) if text is not None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) if "attention_mask" in text_encoding: _UpperCAmelCase = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: _UpperCAmelCase = text_encoding.pop('input_ids' ) else: _UpperCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase ) return encoding_image_processor def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
# flake8: noqa # Lint as: python3 SCREAMING_SNAKE_CASE :str = [ 'VerificationMode', 'Version', 'disable_progress_bar', 'enable_progress_bar', 'is_progress_bar_enabled', 'experimental', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
15
class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase = "" , UpperCAmelCase = False ): """simple docstring""" _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 0 for q, w in zip(self.prefix , UpperCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=UpperCAmelCase , is_leaf=UpperCAmelCase ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def UpperCamelCase ( self , UpperCAmelCase = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = RadixNode() root.insert_many(__lowerCAmelCase ) assert all(root.find(__lowerCAmelCase ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = 'banana bananas bandanas bandana band apple all beast'.split() root.insert_many(__lowerCAmelCase ) print('Words:' , __lowerCAmelCase ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
39
0
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) __lowerCamelCase : List[Any] = logging.getLogger() __lowerCamelCase : List[Any] = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __snake_case ( snake_case__ ): def __a ( self : int , _lowercase : Dict ): """simple docstring""" os.makedirs(_lowercase , exist_ok=_lowercase ) SCREAMING_SNAKE_CASE__ = {"""source""": """What is love ?""", """target""": """life"""} SCREAMING_SNAKE_CASE__ = {"""train""": 12, """val""": 2, """test""": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: SCREAMING_SNAKE_CASE__ = """\n""".join([contents[field]] * n_lines[split] ) with open(os.path.join(_lowercase , f"""{split}.{field}""" ) , """w""" ) as f: f.write(_lowercase ) def __a ( self : Any , _lowercase : Dict , _lowercase : int = "pytorch" ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ = os.path.join(_lowercase , """output""" ) SCREAMING_SNAKE_CASE__ = os.path.join(_lowercase , """data""" ) self._create_dummy_data(data_dir=_lowercase ) SCREAMING_SNAKE_CASE__ = f""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(f"""--gpus={gpus}""" ) if is_apex_available(): testargs.append("""--fp16""" ) else: testargs.append("""--gpus=0""" ) testargs.append("""--distributed_backend=ddp_cpu""" ) testargs.append("""--num_processes=2""" ) SCREAMING_SNAKE_CASE__ = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(_lowercase , env=self.get_env() ) SCREAMING_SNAKE_CASE__ = os.path.join(_lowercase , """metrics.json""" ) with open(_lowercase ) as f: SCREAMING_SNAKE_CASE__ = json.load(_lowercase ) return result @require_torch_gpu def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu def __a ( self : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_gpu @require_ray def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu @require_ray def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 )
219
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class __lowerCamelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments UpperCAmelCase="<s>" , UpperCAmelCase="<pad>" , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = bos, unk, pad, eos _UpperCAmelCase = [] _UpperCAmelCase = [] _UpperCAmelCase = {} _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , UpperCAmelCase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , UpperCAmelCase ): """simple docstring""" return sym in self.indices @classmethod def UpperCamelCase ( cls , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = cls() d.add_from_file(UpperCAmelCase ) return d def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase=1 , UpperCAmelCase=False ): """simple docstring""" if word in self.indices and not overwrite: _UpperCAmelCase = self.indices[word] _UpperCAmelCase = self.count[idx] + n return idx else: _UpperCAmelCase = len(self.symbols ) _UpperCAmelCase = idx self.symbols.append(UpperCAmelCase ) self.count.append(UpperCAmelCase ) return idx def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return 0 def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(UpperCAmelCase ) ) return _UpperCAmelCase = f.readlines() _UpperCAmelCase = self._load_meta(UpperCAmelCase ) for line in lines[indices_start_line:]: try: _UpperCAmelCase , _UpperCAmelCase = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": _UpperCAmelCase = True _UpperCAmelCase , _UpperCAmelCase = line.rsplit(' ' , 1 ) else: _UpperCAmelCase = False _UpperCAmelCase = int(UpperCAmelCase ) _UpperCAmelCase = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(UpperCAmelCase ) ) self.add_symbol(UpperCAmelCase , n=UpperCAmelCase , overwrite=UpperCAmelCase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" _UpperCAmelCase = dict((re.sub(R'@@$' , '' , __lowerCAmelCase ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , __lowerCAmelCase ), v) for k, v in d.items() ) _UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] _UpperCAmelCase = d[k] # restore return da def __A ( __lowerCAmelCase , __lowerCAmelCase )-> str: """simple docstring""" if not os.path.exists(__lowerCAmelCase ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'checkpoint.pt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location='cpu' ) _UpperCAmelCase = chkpt['cfg']['model'] # dicts _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'dict.txt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) _UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) _UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) _UpperCAmelCase = len(__lowerCAmelCase ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['vocab_file'] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # merges_file (bpecodes) _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'bpecodes' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowerCAmelCase , __lowerCAmelCase ) # model config _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'config.json' ) _UpperCAmelCase = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1E-12, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # tokenizer config _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 1_024, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # model _UpperCAmelCase = chkpt['model'] # remove unneeded keys _UpperCAmelCase = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) else: _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase = BioGptConfig.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase = BioGptForCausalLM(__lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(__lowerCAmelCase ) # save _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) print('Conversion is done!' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
39
0
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging lowerCAmelCase__ = logging.get_logger(__name__) class a__ ( snake_case__ ): """simple docstring""" __lowerCamelCase = ['pixel_values'] def __init__( self , lowercase = True , lowercase = 1 / 255 , lowercase = True , lowercase = 8 , **lowercase , ) -> List[str]: '''simple docstring''' super().__init__(**lowercase ) A__ = do_rescale A__ = rescale_factor A__ = do_pad A__ = pad_size def UpperCamelCase ( self , lowercase , lowercase , lowercase = None , **lowercase ) -> int: '''simple docstring''' return rescale(lowercase , scale=lowercase , data_format=lowercase , **lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase = None ) -> Any: '''simple docstring''' A__ , A__ = get_image_size(lowercase ) A__ = (old_height // size + 1) * size - old_height A__ = (old_width // size + 1) * size - old_width return pad(lowercase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowercase ) def UpperCamelCase ( self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = ChannelDimension.FIRST , **lowercase , ) -> Optional[int]: '''simple docstring''' A__ = do_rescale if do_rescale is not None else self.do_rescale A__ = rescale_factor if rescale_factor is not None else self.rescale_factor A__ = do_pad if do_pad is not None else self.do_pad A__ = pad_size if pad_size is not None else self.pad_size A__ = make_list_of_images(lowercase ) if not valid_images(lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. A__ = [to_numpy_array(lowercase ) for image in images] if do_rescale: A__ = [self.rescale(image=lowercase , scale=lowercase ) for image in images] if do_pad: A__ = [self.pad(lowercase , size=lowercase ) for image in images] A__ = [to_channel_dimension_format(lowercase , lowercase ) for image in images] A__ = {"pixel_values": images} return BatchFeature(data=lowercase , tensor_type=lowercase )
68
from __future__ import annotations import collections import pprint from pathlib import Path def __A ( __lowerCAmelCase )-> str: """simple docstring""" return "".join(sorted(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase )-> list[str]: """simple docstring""" return word_by_signature[signature(__lowerCAmelCase )] _a = Path(__file__).parent.joinpath('''words.txt''').read_text(encoding='''utf-8''') _a = sorted({word.strip().lower() for word in data.splitlines()}) _a = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _a = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('''anagrams.txt''', '''w''') as file: file.write('''all_anagrams = \n ''') file.write(pprint.pformat(all_anagrams))
39
0
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig UpperCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase : '''simple docstring''' def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Dict ): """simple docstring""" _A: str = question_encoder _A: Any = generator _A: Dict = self.question_encoder def __magic_name__ ( self : str , lowerCAmelCase_ : Tuple ): """simple docstring""" if os.path.isfile(lowerCAmelCase_ ): raise ValueError(F"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) _A: int = os.path.join(lowerCAmelCase_ , '''question_encoder_tokenizer''' ) _A: Dict = os.path.join(lowerCAmelCase_ , '''generator_tokenizer''' ) self.question_encoder.save_pretrained(lowerCAmelCase_ ) self.generator.save_pretrained(lowerCAmelCase_ ) @classmethod def __magic_name__ ( cls : Optional[int] , lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : Optional[Any] ): """simple docstring""" from ..auto.tokenization_auto import AutoTokenizer _A: Tuple = kwargs.pop('''config''' , lowerCAmelCase_ ) if config is None: _A: List[str] = RagConfig.from_pretrained(lowerCAmelCase_ ) _A: int = AutoTokenizer.from_pretrained( lowerCAmelCase_ , config=config.question_encoder , subfolder='''question_encoder_tokenizer''' ) _A: Optional[Any] = AutoTokenizer.from_pretrained( lowerCAmelCase_ , config=config.generator , subfolder='''generator_tokenizer''' ) return cls(question_encoder=lowerCAmelCase_ , generator=lowerCAmelCase_ ) def __call__( self : int , *lowerCAmelCase_ : Optional[Any] , **lowerCAmelCase_ : str ): """simple docstring""" return self.current_tokenizer(*lowerCAmelCase_ , **lowerCAmelCase_ ) def __magic_name__ ( self : str , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : Tuple ): """simple docstring""" return self.generator.batch_decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) def __magic_name__ ( self : Tuple , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Optional[int] ): """simple docstring""" return self.generator.decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) def __magic_name__ ( self : List[str] ): """simple docstring""" _A: str = self.question_encoder def __magic_name__ ( self : Tuple ): """simple docstring""" _A: str = self.generator def __magic_name__ ( self : Any , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : int = None , lowerCAmelCase_ : Any = None , lowerCAmelCase_ : List[str] = None , lowerCAmelCase_ : int = "longest" , lowerCAmelCase_ : Union[str, Any] = None , lowerCAmelCase_ : str = True , **lowerCAmelCase_ : List[Any] , ): """simple docstring""" warnings.warn( '''`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ''' '''regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ''' '''context manager to prepare your targets. See the documentation of your specific tokenizer for more ''' '''details''' , lowerCAmelCase_ , ) if max_length is None: _A: Any = self.current_tokenizer.model_max_length _A: List[str] = self( lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , max_length=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , **lowerCAmelCase_ , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: _A: List[str] = self.current_tokenizer.model_max_length _A: Union[str, Any] = self( text_target=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , padding=lowerCAmelCase_ , max_length=lowerCAmelCase_ , truncation=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A: List[Any] = labels['''input_ids'''] return model_inputs
121
from __future__ import annotations def __A ( __lowerCAmelCase )-> list[int]: """simple docstring""" _UpperCAmelCase = 2 _UpperCAmelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCAmelCase ) if n > 1: factors.append(__lowerCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
39
0
"""simple docstring""" from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class _UpperCAmelCase( snake_case__ ): lowercase__ = DistilBertTokenizer lowercase__ = DistilBertTokenizerFast lowercase__ = True @slow def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = DistilBertTokenizer.from_pretrained('''distilbert-base-uncased''') _UpperCamelCase = tokenizer.encode('''sequence builders''' , add_special_tokens=__a) _UpperCamelCase = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__a) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a , __a) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
194
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __A ( )-> tuple[list[int], int]: """simple docstring""" _UpperCAmelCase = [randint(-1_000 , 1_000 ) for i in range(10 )] _UpperCAmelCase = randint(-5_000 , 5_000 ) return (arr, r) _a = make_dataset() def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, ...]: """simple docstring""" for triplet in permutations(__lowerCAmelCase , 3 ): if sum(__lowerCAmelCase ) == target: return tuple(sorted(__lowerCAmelCase ) ) return (0, 0, 0) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, int, int]: """simple docstring""" arr.sort() _UpperCAmelCase = len(__lowerCAmelCase ) for i in range(n - 1 ): _UpperCAmelCase , _UpperCAmelCase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __A ( )-> tuple[float, float]: """simple docstring""" _UpperCAmelCase = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' _UpperCAmelCase = '\ntriplet_sum1(*dataset)\n' _UpperCAmelCase = '\ntriplet_sum2(*dataset)\n' _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) return (min(__lowerCAmelCase ), min(__lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _a = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
39
0
def lowerCamelCase_ ( _UpperCamelCase ) -> list[int]: """simple docstring""" snake_case_ : Union[str, Any] = len(__lowerCAmelCase ) for i in range(__lowerCAmelCase ): for j in range(i + 1 , __lowerCAmelCase ): if numbers[j] < numbers[i]: snake_case_ , snake_case_ : List[str] = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowerCAmelCase_ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase_ = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
279
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase , cache_dir=UpperCAmelCase ) _UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(UpperCAmelCase , os.listdir(UpperCAmelCase )[0] , 'snapshots' ) )] _UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 4 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1e-3 assert np.abs(np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5e-1 _UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(UpperCAmelCase ) == num_samples def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , set_alpha_to_one=UpperCAmelCase , steps_offset=1 , ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=UpperCAmelCase , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = scheduler.create_state() _UpperCAmelCase = scheduler_state _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # With memory efficient attention _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , use_memory_efficient_attention=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
39
0
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''' , [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 1_0, '''max_num_jobs''': 1}, [range(1_0 )]), ({'''num_shards''': 1_0, '''max_num_jobs''': 1_0}, [range(__lowerCAmelCase , i + 1 ) for i in range(1_0 )]), ({'''num_shards''': 1, '''max_num_jobs''': 1_0}, [range(1 )]), ({'''num_shards''': 1_0, '''max_num_jobs''': 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 1_0 )]), ({'''num_shards''': 3, '''max_num_jobs''': 1_0}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : int , lowerCAmelCase__ : Tuple ): __a : Tuple = _distribute_shards(**__lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''' , [ ({'''foo''': 0}, 1_0, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : int ): __a : Dict = _split_gen_kwargs(__lowerCAmelCase , __lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''' , [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Any ): if expected is RuntimeError: with pytest.raises(__lowerCAmelCase ): _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) else: __a : List[str] = _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) assert out == expected
216
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = AlbertTokenizer UpperCamelCase__ = AlbertTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = True def UpperCamelCase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 'this is a test' _UpperCAmelCase = 'this is a test' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = '<pad>' _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(UpperCAmelCase ) , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = 'I was born in 92000, and this is falsé.' _UpperCAmelCase = tokenizer.tokenize(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [48, 25, 21, 1289] ) _UpperCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) _UpperCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode('sequence builders' ) _UpperCAmelCase = tokenizer.encode('multi-sequence build' ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
39
0
import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( "The `image_to_image.py` script is outdated. Please use directly `from diffusers import" " StableDiffusionImg2ImgPipeline` instead." )
218
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): lowerCamelCase__: Optional[int] = StableDiffusionXLImgaImgPipeline lowerCamelCase__: Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} lowerCamelCase__: str = PipelineTesterMixin.required_optional_params - {"latents"} lowerCamelCase__: Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCamelCase__: Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__: List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def _lowerCamelCase ( self: Union[str, Any] ) -> int: torch.manual_seed(0 ) __UpperCAmelCase : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , attention_head_dim=(2, 4) , use_linear_projection=__lowerCamelCase , addition_embed_type="text_time" , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __UpperCAmelCase : List[Any] = EulerDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , steps_offset=1 , beta_schedule="scaled_linear" , timestep_spacing="leading" , ) torch.manual_seed(0 ) __UpperCAmelCase : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) __UpperCAmelCase : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=32 , ) __UpperCAmelCase : List[str] = CLIPTextModel(__lowerCamelCase ) __UpperCAmelCase : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=__lowerCamelCase ) __UpperCAmelCase : str = CLIPTextModelWithProjection(__lowerCamelCase ) __UpperCAmelCase : int = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=__lowerCamelCase ) __UpperCAmelCase : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_encoder_2": text_encoder_a, "tokenizer_2": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def _lowerCamelCase ( self: Union[str, Any] , __lowerCamelCase: List[Any] , __lowerCamelCase: List[Any]=0 ) -> Optional[int]: __UpperCAmelCase : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = image / 2 + 0.5 if str(__lowerCamelCase ).startswith("mps" ): __UpperCAmelCase : str = torch.manual_seed(__lowerCamelCase ) else: __UpperCAmelCase : Dict = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) __UpperCAmelCase : str = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 5.0, "output_type": "numpy", "strength": 0.75, } return inputs def _lowerCamelCase ( self: Tuple ) -> Tuple: __UpperCAmelCase : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator __UpperCAmelCase : List[str] = self.get_dummy_components() __UpperCAmelCase : int = StableDiffusionXLImgaImgPipeline(**__lowerCamelCase ) __UpperCAmelCase : Any = sd_pipe.to(__lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCamelCase ) __UpperCAmelCase : Tuple = self.get_dummy_inputs(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = sd_pipe(**__lowerCamelCase ).images __UpperCAmelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCAmelCase : Union[str, Any] = np.array([0.46_56, 0.48_40, 0.44_39, 0.66_98, 0.55_74, 0.45_24, 0.57_99, 0.59_43, 0.51_65] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowerCamelCase ( self: int ) -> Dict: super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def _lowerCamelCase ( self: Optional[int] ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def _lowerCamelCase ( self: str ) -> int: pass def _lowerCamelCase ( self: Tuple ) -> Dict: __UpperCAmelCase : str = self.get_dummy_components() __UpperCAmelCase : Dict = StableDiffusionXLImgaImgPipeline(**__lowerCamelCase ) __UpperCAmelCase : List[str] = sd_pipe.to(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = sd_pipe.to(__lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCamelCase ) # forward without prompt embeds __UpperCAmelCase : int = self.get_dummy_inputs(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = 3 * ["this is a negative prompt"] __UpperCAmelCase : List[Any] = negative_prompt __UpperCAmelCase : Optional[int] = 3 * [inputs["prompt"]] __UpperCAmelCase : Optional[int] = sd_pipe(**__lowerCamelCase ) __UpperCAmelCase : Dict = output.images[0, -3:, -3:, -1] # forward with prompt embeds __UpperCAmelCase : Tuple = self.get_dummy_inputs(__lowerCamelCase ) __UpperCAmelCase : Any = 3 * ["this is a negative prompt"] __UpperCAmelCase : List[Any] = 3 * [inputs.pop("prompt" )] ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) : Tuple = sd_pipe.encode_prompt(__lowerCamelCase , negative_prompt=__lowerCamelCase ) __UpperCAmelCase : List[str] = sd_pipe( **__lowerCamelCase , prompt_embeds=__lowerCamelCase , negative_prompt_embeds=__lowerCamelCase , pooled_prompt_embeds=__lowerCamelCase , negative_pooled_prompt_embeds=__lowerCamelCase , ) __UpperCAmelCase : Any = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class _snake_case ( unittest.TestCase ): def _lowerCamelCase ( self: Any ) -> List[str]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self: Dict , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: int="cpu" , __lowerCamelCase: Dict=torch.floataa , __lowerCamelCase: List[str]=0 ) -> int: __UpperCAmelCase : int = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) __UpperCAmelCase : List[str] = np.random.RandomState(__lowerCamelCase ).standard_normal((1, 4, 64, 64) ) __UpperCAmelCase : Dict = torch.from_numpy(__lowerCamelCase ).to(device=__lowerCamelCase , dtype=__lowerCamelCase ) __UpperCAmelCase : str = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def _lowerCamelCase ( self: Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase : Any = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-base" ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = self.get_inputs(__lowerCamelCase ) __UpperCAmelCase : Dict = pipe(**__lowerCamelCase ).images __UpperCAmelCase : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __UpperCAmelCase : List[str] = np.array([0.4_94_93, 0.4_78_96, 0.4_07_98, 0.5_42_14, 0.5_32_12, 0.4_82_02, 0.4_76_56, 0.4_63_29, 0.4_85_06] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
157
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "distilbert" UpperCamelCase__ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , UpperCAmelCase=3_0522 , UpperCAmelCase=512 , UpperCAmelCase=False , UpperCAmelCase=6 , UpperCAmelCase=12 , UpperCAmelCase=768 , UpperCAmelCase=4 * 768 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=0.02 , UpperCAmelCase=0.1 , UpperCAmelCase=0.2 , UpperCAmelCase=0 , **UpperCAmelCase , ): """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = sinusoidal_pos_embds _UpperCAmelCase = n_layers _UpperCAmelCase = n_heads _UpperCAmelCase = dim _UpperCAmelCase = hidden_dim _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation _UpperCAmelCase = initializer_range _UpperCAmelCase = qa_dropout _UpperCAmelCase = seq_classif_dropout super().__init__(**UpperCAmelCase , pad_token_id=UpperCAmelCase ) class __lowerCamelCase ( snake_case__): """simple docstring""" @property def UpperCamelCase ( self ): """simple docstring""" if self.task == "multiple-choice": _UpperCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _UpperCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
39
0
'''simple docstring''' def __a(SCREAMING_SNAKE_CASE_ : Optional[Any] = 10**12 ): '''simple docstring''' _lowerCAmelCase = 1 _lowerCAmelCase = 0 _lowerCAmelCase = 1 _lowerCAmelCase = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'''{solution() = }''')
158
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() _a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCamelCase ( snake_case__): """simple docstring""" def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {'source': 'What is love ?', 'target': 'life'} _UpperCAmelCase = {'train': 12, 'val': 2, 'test': 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: _UpperCAmelCase = '\n'.join([contents[field]] * n_lines[split] ) with open(os.path.join(UpperCAmelCase , F"""{split}.{field}""" ) , 'w' ) as f: f.write(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase = "pytorch" ): """simple docstring""" _UpperCAmelCase = self.get_auto_remove_tmp_dir() _UpperCAmelCase = os.path.join(UpperCAmelCase , 'output' ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'data' ) self._create_dummy_data(data_dir=UpperCAmelCase ) _UpperCAmelCase = F""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(F"""--gpus={gpus}""" ) if is_apex_available(): testargs.append('--fp16' ) else: testargs.append('--gpus=0' ) testargs.append('--distributed_backend=ddp_cpu' ) testargs.append('--num_processes=2' ) _UpperCAmelCase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(UpperCAmelCase , env=self.get_env() ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'metrics.json' ) with open(UpperCAmelCase ) as f: _UpperCAmelCase = json.load(UpperCAmelCase ) return result @require_torch_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 )
39
0
import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel lowerCamelCase__ = """0.12""" # assumed parallelism: 8 @require_flax @is_staging_test class A__ ( unittest.TestCase ): @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = TOKEN HfFolder.save_token(a ) @classmethod def _lowerCamelCase ( cls : Tuple ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-model-flax' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-model-flax-org' ) except HTTPError: pass def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) lowerCAmelCase__ : Optional[Any] = FlaxBertModel(a ) model.push_to_hub('test-model-flax' , use_auth_token=self._token ) lowerCAmelCase__ : Any = FlaxBertModel.from_pretrained(f'''{USER}/test-model-flax''' ) lowerCAmelCase__ : Any = flatten_dict(unfreeze(model.params ) ) lowerCAmelCase__ : Union[str, Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): lowerCAmelCase__ : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(a , 1E-3 , msg=f'''{key} not identical''' ) # Reset repo delete_repo(token=self._token , repo_id='test-model-flax' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(a , repo_id='test-model-flax' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : int = FlaxBertModel.from_pretrained(f'''{USER}/test-model-flax''' ) lowerCAmelCase__ : Tuple = flatten_dict(unfreeze(model.params ) ) lowerCAmelCase__ : List[str] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): lowerCAmelCase__ : Optional[int] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(a , 1E-3 , msg=f'''{key} not identical''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) lowerCAmelCase__ : Dict = FlaxBertModel(a ) model.push_to_hub('valid_org/test-model-flax-org' , use_auth_token=self._token ) lowerCAmelCase__ : Optional[int] = FlaxBertModel.from_pretrained('valid_org/test-model-flax-org' ) lowerCAmelCase__ : Optional[Any] = flatten_dict(unfreeze(model.params ) ) lowerCAmelCase__ : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): lowerCAmelCase__ : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(a , 1E-3 , msg=f'''{key} not identical''' ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-model-flax-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( a , repo_id='valid_org/test-model-flax-org' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : Optional[int] = FlaxBertModel.from_pretrained('valid_org/test-model-flax-org' ) lowerCAmelCase__ : List[str] = flatten_dict(unfreeze(model.params ) ) lowerCAmelCase__ : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): lowerCAmelCase__ : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(a , 1E-3 , msg=f'''{key} not identical''' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[Any]: lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Any = flatten_dict(modela.params ) lowerCAmelCase__ : List[Any] = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: lowerCAmelCase__ : List[str] = False return models_are_equal @require_flax class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = BertConfig.from_pretrained('hf-internal-testing/tiny-bert-flax-only' ) lowerCAmelCase__ : List[Any] = FlaxBertModel(a ) lowerCAmelCase__ : Optional[Any] = 'bert' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(a , a ) ) with self.assertRaises(a ): lowerCAmelCase__ : List[Any] = FlaxBertModel.from_pretrained(a ) lowerCAmelCase__ : Any = FlaxBertModel.from_pretrained(a , subfolder=a ) self.assertTrue(check_models_equal(a , a ) ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-bert-flax-only' ) lowerCAmelCase__ : Optional[int] = FlaxBertModel(a ) lowerCAmelCase__ : Dict = 'bert' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(a , a ) , max_shard_size='10KB' ) with self.assertRaises(a ): lowerCAmelCase__ : List[Any] = FlaxBertModel.from_pretrained(a ) lowerCAmelCase__ : str = FlaxBertModel.from_pretrained(a , subfolder=a ) self.assertTrue(check_models_equal(a , a ) ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'bert' lowerCAmelCase__ : Union[str, Any] = 'hf-internal-testing/tiny-random-bert-subfolder' with self.assertRaises(a ): lowerCAmelCase__ : Optional[int] = FlaxBertModel.from_pretrained(a ) lowerCAmelCase__ : Union[str, Any] = FlaxBertModel.from_pretrained(a , subfolder=a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'bert' lowerCAmelCase__ : int = 'hf-internal-testing/tiny-random-bert-sharded-subfolder' with self.assertRaises(a ): lowerCAmelCase__ : Dict = FlaxBertModel.from_pretrained(a ) lowerCAmelCase__ : Dict = FlaxBertModel.from_pretrained(a , subfolder=a ) self.assertIsNotNone(a )
212
class __lowerCamelCase : """simple docstring""" def __init__( self ): """simple docstring""" _UpperCAmelCase = {} # Mapping from char to TrieNode _UpperCAmelCase = False def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: _UpperCAmelCase = TrieNode() _UpperCAmelCase = curr.nodes[char] _UpperCAmelCase = True def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: return False _UpperCAmelCase = curr.nodes[char] return curr.is_leaf def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" def _delete(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: if index == len(UpperCAmelCase ): # If word does not exist if not curr.is_leaf: return False _UpperCAmelCase = False return len(curr.nodes ) == 0 _UpperCAmelCase = word[index] _UpperCAmelCase = curr.nodes.get(UpperCAmelCase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted _UpperCAmelCase = _delete(UpperCAmelCase , UpperCAmelCase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCAmelCase , 0 ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" if node.is_leaf: print(__lowerCAmelCase , end=' ' ) for key, value in node.nodes.items(): print_words(__lowerCAmelCase , word + key ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = TrieNode() root.insert_many(__lowerCAmelCase ) # print_words(root, "") assert all(root.find(__lowerCAmelCase ) for word in words ) assert root.find('banana' ) assert not root.find('bandanas' ) assert not root.find('apps' ) assert root.find('apple' ) assert root.find('all' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" print(str(__lowerCAmelCase ) , 'works!' if passes else 'doesn\'t work :(' ) def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" print_results('Testing trie functionality' , test_trie() ) if __name__ == "__main__": main()
39
0
import random from typing import Any def UpperCAmelCase ( a_ ) -> list[Any]: """simple docstring""" for _ in range(len(__lowerCAmelCase ) ): __A = random.randint(0 , len(__lowerCAmelCase ) - 1 ) __A = random.randint(0 , len(__lowerCAmelCase ) - 1 ) __A , __A = data[b], data[a] return data if __name__ == "__main__": SCREAMING_SNAKE_CASE :List[Any] = [0, 1, 2, 3, 4, 5, 6, 7] SCREAMING_SNAKE_CASE :Union[str, Any] = ['python', 'says', 'hello', '!'] print('Fisher-Yates Shuffle:') print('List', integers, strings) print('FY Shuffle', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
15
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _a = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class __lowerCamelCase ( unittest.TestCase): """simple docstring""" UpperCamelCase__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = ZeroShotClassificationPipeline( model=UpperCAmelCase , tokenizer=UpperCAmelCase , candidate_labels=['polics', 'health'] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # No kwarg _UpperCAmelCase = classifier('Who are you voting for in 2020?' , ['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics, public health' ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics', 'public health'] ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='This text is about {}' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # https://github.com/huggingface/transformers/issues/13846 _UpperCAmelCase = classifier(['I am happy'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(1 ) ] , ) _UpperCAmelCase = classifier(['I am happy', 'I am sad'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(2 ) ] , ) with self.assertRaises(UpperCAmelCase ): classifier('' , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier(UpperCAmelCase , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels='' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='Not formatting template' , ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template=UpperCAmelCase , ) self.run_entailment_id(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = zero_shot_classifier.model.config _UpperCAmelCase = config.labelaid _UpperCAmelCase = zero_shot_classifier.entailment_id _UpperCAmelCase = {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) _UpperCAmelCase = {'entailment': 0, 'neutral': 1, 'contradiction': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 0, 'NON-ENTAIL': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 2, 'NEUTRAL': 1, 'CONTR': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) _UpperCAmelCase = original_labelaid self.assertEqual(UpperCAmelCase , zero_shot_classifier.entailment_id ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( 'Who are you voting for in 2020?' * 100 , candidate_labels=['politics', 'public health', 'science'] ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='tf' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @slow @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='pt' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , ) @slow @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='tf' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , )
39
0
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __lowerCamelCase : Any = datasets.logging.get_logger(__name__) __lowerCamelCase : Any = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' __lowerCamelCase : Union[str, Any] = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' __lowerCamelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Union[str, Any]=False , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : Tuple=True , __UpperCamelCase : Tuple=False , __UpperCamelCase : Any="dummy_doc" ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = {doc: key_lines} SCREAMING_SNAKE_CASE__ = {doc: sys_lines} SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = reader.get_doc_mentions(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: SCREAMING_SNAKE_CASE__ = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = reader.get_doc_mentions(__lowerCAmelCase , sys_doc_lines[doc] , __lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: SCREAMING_SNAKE_CASE__ = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) if remove_nested: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters SCREAMING_SNAKE_CASE__ = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( """Number of resulting singleton clusters in the key """ f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ """files, respectively""" ) return doc_coref_infos def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Dict , __UpperCamelCase : Tuple , __UpperCamelCase : Any , __UpperCamelCase : Tuple , __UpperCamelCase : List[str] , __UpperCamelCase : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_coref_infos(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 for name, metric in metrics: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = evaluator.evaluate_documents(__lowerCAmelCase , __lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , f"""Recall: {recall * 1_00:.2f}""" , f""" Precision: {precision * 1_00:.2f}""" , f""" F1: {fa * 1_00:.2f}""" , ) if conll_subparts_num == 3: SCREAMING_SNAKE_CASE__ = (conll / 3) * 1_00 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({"""conll_score""": conll} ) return output_scores def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Tuple ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: SCREAMING_SNAKE_CASE__ = line.split()[5] if not parse_col == "-": SCREAMING_SNAKE_CASE__ = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): def __a ( self : str ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def __a ( self : List[Any] , _lowercase : str , _lowercase : Optional[int] , _lowercase : List[str]=True , _lowercase : List[str]=False , _lowercase : List[str]=False , _lowercase : Optional[int]=False ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: SCREAMING_SNAKE_CASE__ = util.check_gold_parse_annotation(_lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use \'min_span\'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" SCREAMING_SNAKE_CASE__ = evaluate( key_lines=_lowercase , sys_lines=_lowercase , metrics=_lowercase , NP_only=_lowercase , remove_nested=_lowercase , keep_singletons=_lowercase , min_span=_lowercase , ) return score
219
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _a = get_logger(__name__) class __lowerCamelCase ( enum.Enum): """simple docstring""" UpperCamelCase__ = "all_checks" UpperCamelCase__ = "basic_checks" UpperCamelCase__ = "no_checks" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None )-> str: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase = ' for ' + verification_name if verification_name is not None else '' if len(__lowerCAmelCase ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__lowerCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__lowerCAmelCase ) ) logger.info('All the splits matched successfully.' ) def __A ( __lowerCAmelCase , __lowerCAmelCase = True )-> dict: """simple docstring""" if record_checksum: _UpperCAmelCase = shaaaa() with open(__lowerCAmelCase , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__lowerCAmelCase ) _UpperCAmelCase = m.hexdigest() else: _UpperCAmelCase = None return {"num_bytes": os.path.getsize(__lowerCAmelCase ), "checksum": checksum} def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
39
0
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
68
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=16 , UpperCAmelCase=[1, 2, 1] , UpperCAmelCase=[2, 2, 4] , UpperCAmelCase=2 , UpperCAmelCase=2.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=10 , UpperCAmelCase=8 , UpperCAmelCase=["stage1", "stage2", "stage3"] , UpperCAmelCase=[1, 2, 3] , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = embed_dim _UpperCAmelCase = depths _UpperCAmelCase = num_heads _UpperCAmelCase = window_size _UpperCAmelCase = mlp_ratio _UpperCAmelCase = qkv_bias _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = drop_path_rate _UpperCAmelCase = hidden_act _UpperCAmelCase = use_absolute_embeddings _UpperCAmelCase = patch_norm _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = initializer_range _UpperCAmelCase = is_training _UpperCAmelCase = scope _UpperCAmelCase = use_labels _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = encoder_stride _UpperCAmelCase = out_features _UpperCAmelCase = out_indices def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) _UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCAmelCase ): _UpperCAmelCase = ['stem'] _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCamelCase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) # Swin has a different seq_length _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = 3 _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCAmelCase ): _UpperCAmelCase = 0 return t def check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase={} ): with torch.no_grad(): _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ).to_tuple() def recursive_check(UpperCAmelCase , UpperCAmelCase ): if isinstance(UpperCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase , UpperCAmelCase ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCAmelCase ) , set_nan_tensor_to_zero(UpperCAmelCase ) , atol=1e-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}. Dict has""" F""" `nan`: {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}.""" ) , ) recursive_check(UpperCAmelCase , UpperCAmelCase ) for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) @require_torch class __lowerCamelCase ( unittest.TestCase , snake_case__): """simple docstring""" UpperCamelCase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCamelCase__ = MaskFormerSwinConfig def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: _UpperCAmelCase = backbone_class(UpperCAmelCase ) backbone.to(UpperCAmelCase ) backbone.eval() _UpperCAmelCase = backbone(**UpperCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _UpperCAmelCase = backbone(**UpperCAmelCase , output_hidden_states=UpperCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _UpperCAmelCase = backbone(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertIsNotNone(outputs.attentions )
39
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase__ : int = logging.get_logger(__name__) def lowerCamelCase__ ( a , a=False ) -> Union[str, Any]: _A: List[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _A: Tuple = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( a , a , a=False ) -> List[str]: for i in range(config.num_hidden_layers ): if base_model: _A: Optional[Any] = '''''' else: _A: Dict = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _A: Any = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) _A: Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _A: Optional[Any] = in_proj_weight[ : config.hidden_size, : ] _A: List[Any] = in_proj_bias[: config.hidden_size] _A: List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _A: Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _A: List[str] = in_proj_weight[ -config.hidden_size :, : ] _A: Union[str, Any] = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( a ) -> Optional[Any]: _A: Any = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def lowerCamelCase__ ( a , a , a ) -> int: _A: List[Any] = dct.pop(__lowerCAmelCase ) _A: Optional[Any] = val def lowerCamelCase__ ( ) -> str: _A: Union[str, Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _A: Dict = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( a , a , a=True ) -> List[str]: _A: List[str] = ViTConfig() # patch_size if model_name[-1] == "8": _A: Tuple = 8 # set labels if required if not base_model: _A: int = 10_00 _A: Any = '''huggingface/label-files''' _A: Optional[int] = '''imagenet-1k-id2label.json''' _A: Optional[int] = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) _A: Tuple = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _A: int = idalabel _A: Any = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _A: Any = 3_84 _A: Any = 15_36 _A: Dict = 12 _A: Any = 6 # load original model from torch hub _A: Optional[Any] = torch.hub.load('''facebookresearch/dino:main''' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _A: Tuple = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _A: str = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _A: Tuple = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _A: List[Any] = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _A: List[str] = ViTImageProcessor() _A: int = image_processor(images=prepare_img() , return_tensors='''pt''' ) _A: Optional[Any] = encoding['''pixel_values'''] _A: Dict = model(__lowerCAmelCase ) if base_model: _A: Any = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _A: Any = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) UpperCAmelCase__ : Union[str, Any] = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
121
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = TransfoXLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" super().setUp() _UpperCAmelCase = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = '<unk> UNwanted , running' _UpperCAmelCase = '<unk> unwanted, running' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(UpperCAmelCase , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [0, 4, 8, 7] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) _UpperCAmelCase = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' _UpperCAmelCase = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(UpperCAmelCase ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(UpperCAmelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
39
0
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" if not head: return True # split the list to two parts _UpperCamelCase , _UpperCamelCase = head.next, head while fast and fast.next: _UpperCamelCase = fast.next.next _UpperCamelCase = slow.next _UpperCamelCase = slow.next _UpperCamelCase = None # Don't forget here! But forget still works! # reverse the second part _UpperCamelCase = None while second: _UpperCamelCase = second.next _UpperCamelCase = node _UpperCamelCase = second _UpperCamelCase = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False _UpperCamelCase = node.next _UpperCamelCase = head.next return True def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" if not head or not head.next: return True # 1. Get the midpoint (slow) _UpperCamelCase = _UpperCamelCase = _UpperCamelCase = head while fast and fast.next: _UpperCamelCase , _UpperCamelCase = fast.next.next, slow.next # 2. Push the second half into the stack _UpperCamelCase = [slow.val] while slow.next: _UpperCamelCase = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False _UpperCamelCase = cur.next return True def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" if not head or not head.next: return True _UpperCamelCase = {} _UpperCamelCase = 0 while head: if head.val in d: d[head.val].append(__lowerCAmelCase ) else: _UpperCamelCase = [pos] _UpperCamelCase = head.next pos += 1 _UpperCamelCase = pos - 1 _UpperCamelCase = 0 for v in d.values(): if len(__lowerCAmelCase ) % 2 != 0: middle += 1 else: _UpperCamelCase = 0 for i in range(0, len(__lowerCAmelCase ) ): if v[i] + v[len(__lowerCAmelCase ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
194
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = {'''configuration_opt''': ['''OPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OPTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''OPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OPTForCausalLM''', '''OPTModel''', '''OPTPreTrainedModel''', '''OPTForSequenceClassification''', '''OPTForQuestionAnswering''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['''TFOPTForCausalLM''', '''TFOPTModel''', '''TFOPTPreTrainedModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''FlaxOPTForCausalLM''', '''FlaxOPTModel''', '''FlaxOPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
39
0
def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" snake_case_ : List[str] = abs(__lowerCAmelCase ) snake_case_ : Any = 0 while n > 0: res += n % 10 n //= 10 return res def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" snake_case_ : List[str] = abs(__lowerCAmelCase ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" return sum(int(__lowerCAmelCase ) for c in str(abs(__lowerCAmelCase ) ) ) def lowerCamelCase_ ( ) -> None: """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(_UpperCamelCase , _UpperCamelCase ) -> None: snake_case_ : Union[str, Any] = f'''{func.__name__}({value})''' snake_case_ : Union[str, Any] = timeit(f'''__main__.{call}''' , setup='''import __main__''' ) print(f'''{call:56} = {func(__lowerCAmelCase )} -- {timing:.4f} seconds''' ) for value in (262_144, 1_125_899_906_842_624, 1_267_650_600_228_229_401_496_703_205_376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(__lowerCAmelCase , __lowerCAmelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
279
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" _UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ), F"""{len(__lowerCAmelCase )} != {len(__lowerCAmelCase )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) _a = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _a = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def __A ( __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" try: _UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(__lowerCAmelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __A ( __lowerCAmelCase , __lowerCAmelCase = "student" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , )-> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" _UpperCAmelCase = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowerCAmelCase , __lowerCAmelCase ): AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) # purely for convenience _UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(__lowerCAmelCase ).eval() else: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), F"""teacher must be a model or string got type {type(__lowerCAmelCase )}""" _UpperCAmelCase = teacher.config.to_diff_dict() try: _UpperCAmelCase , _UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowerCAmelCase ) # Copy weights _UpperCAmelCase = teacher.config_class(**__lowerCAmelCase ) _UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. _UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=__lowerCAmelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save _UpperCAmelCase , _UpperCAmelCase = list(range(__lowerCAmelCase ) ), list(range(__lowerCAmelCase ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(__lowerCAmelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) if d_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) try: if hasattr( __lowerCAmelCase , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowerCAmelCase ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowerCAmelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowerCAmelCase ) copy_layers(teacher.decoder.block , student.decoder.block , __lowerCAmelCase ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) _UpperCAmelCase = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowerCAmelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
39
0
from __future__ import annotations class UpperCamelCase__ : def __init__(self : Any , snake_case_ : List[str] , snake_case_ : Union[str, Any] ): __a , __a : int = text, pattern __a , __a : Dict = len(snake_case_ ), len(snake_case_ ) def lowerCAmelCase (self : List[str] , snake_case_ : List[str] ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCAmelCase (self : Optional[Any] , snake_case_ : Tuple ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCAmelCase (self : int ): __a : Any = [] for i in range(self.textLen - self.patLen + 1 ): __a : Optional[Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: __a : Optional[int] = self.match_in_pattern(self.text[mismatch_index] ) __a : int = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions lowercase__ ='ABAABA' lowercase__ ='AB' lowercase__ =BoyerMooreSearch(text, pattern) lowercase__ =bms.bad_character_heuristic() if len(positions) == 0: print('No match found') else: print('Pattern found in following positions: ') print(positions)
216
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False )-> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase = '' else: _UpperCAmelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase = in_proj_bias[: config.hidden_size] _UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase = in_proj_bias[-config.hidden_size :] def __A ( __lowerCAmelCase )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" _UpperCAmelCase = dct.pop(__lowerCAmelCase ) _UpperCAmelCase = val def __A ( )-> str: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True )-> List[str]: """simple docstring""" _UpperCAmelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCAmelCase = 8 # set labels if required if not base_model: _UpperCAmelCase = 1_000 _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCAmelCase = 384 _UpperCAmelCase = 1_536 _UpperCAmelCase = 12 _UpperCAmelCase = 6 # load original model from torch hub _UpperCAmelCase = torch.hub.load('facebookresearch/dino:main' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCAmelCase = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _UpperCAmelCase = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _UpperCAmelCase = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCAmelCase = ViTImageProcessor() _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCAmelCase = encoding['pixel_values'] _UpperCAmelCase = model(__lowerCAmelCase ) if base_model: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) _a = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
39
0
import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def UpperCamelCase_( _snake_case : Optional[int] ): """simple docstring""" __a =[] if isinstance(__lowerCAmelCase , __lowerCAmelCase ): for v in tree.values(): shapes.extend(_fetch_dims(__lowerCAmelCase ) ) elif isinstance(__lowerCAmelCase , (list, tuple) ): for t in tree: shapes.extend(_fetch_dims(__lowerCAmelCase ) ) elif isinstance(__lowerCAmelCase , torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError('Not supported' ) return shapes @torch.jit.ignore def UpperCamelCase_( _snake_case : Any , _snake_case : Dict ): """simple docstring""" __a =[] for d in reversed(__lowerCAmelCase ): idx.append(flat_idx % d ) __a =flat_idx // d return tuple(reversed(__lowerCAmelCase ) ) @torch.jit.ignore def UpperCamelCase_( _snake_case : Tuple , _snake_case : str , _snake_case : int , _snake_case : Optional[Any] = None , _snake_case : str = None , ): """simple docstring""" def reduce_edge_list(_snake_case : str ) -> None: __a =True for i in range(len(__lowerCAmelCase ) ): __a =-1 * (i + 1) l[reversed_idx] &= tally __a =l[reversed_idx] if start_edges is None: __a =[s == 0 for s in start] reduce_edge_list(__lowerCAmelCase ) if end_edges is None: __a =[e == (d - 1) for e, d in zip(__lowerCAmelCase , __lowerCAmelCase )] reduce_edge_list(__lowerCAmelCase ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(__lowerCAmelCase ) == 0: return [()] elif len(__lowerCAmelCase ) == 1: return [(slice(start[0] , end[0] + 1 ),)] __a =[] __a =[] # Dimensions common to start and end can be selected directly for s, e in zip(__lowerCAmelCase , __lowerCAmelCase ): if s == e: path_list.append(slice(__lowerCAmelCase , s + 1 ) ) else: break __a =tuple(__lowerCAmelCase ) __a =len(__lowerCAmelCase ) # start == end, and we're done if divergence_idx == len(__lowerCAmelCase ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __a =start[divergence_idx] return tuple( path + (slice(__lowerCAmelCase , sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] , [d - 1 for d in dims[divergence_idx + 1 :]] , dims[divergence_idx + 1 :] , start_edges=start_edges[divergence_idx + 1 :] , end_edges=[True for _ in end_edges[divergence_idx + 1 :]] , ) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __a =end[divergence_idx] return tuple( path + (slice(__lowerCAmelCase , edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] , end[divergence_idx + 1 :] , dims[divergence_idx + 1 :] , start_edges=[True for _ in start_edges[divergence_idx + 1 :]] , end_edges=end_edges[divergence_idx + 1 :] , ) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) __a =end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def UpperCamelCase_( _snake_case : List[str] , _snake_case : List[str] , _snake_case : List[Any] , _snake_case : List[Any] ): """simple docstring""" __a =t.shape[:no_batch_dims] __a =list(_flat_idx_to_idx(__lowerCAmelCase , __lowerCAmelCase ) ) # _get_minimal_slice_set is inclusive __a =list(_flat_idx_to_idx(flat_end - 1 , __lowerCAmelCase ) ) # Get an ordered list of slices to perform __a =_get_minimal_slice_set( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) __a =[t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def UpperCamelCase_( _snake_case : Optional[Any] , _snake_case : Optional[Any] , _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : Tuple = False , _snake_case : Optional[int] = None , _snake_case : int = False , ): """simple docstring""" if not (len(__lowerCAmelCase ) > 0): raise ValueError('Must provide at least one input' ) __a =[shape[:no_batch_dims] for shape in _fetch_dims(__lowerCAmelCase )] __a =tuple([max(__lowerCAmelCase ) for s in zip(*__lowerCAmelCase )] ) def _prep_inputs(_snake_case : Optional[int] ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: __a =t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) __a =t.reshape(-1 , *t.shape[no_batch_dims:] ) else: __a =t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t __a =tensor_tree_map(_prep_inputs , __lowerCAmelCase ) __a =None if _out is not None: __a =tensor_tree_map(lambda _snake_case : t.view([-1] + list(t.shape[no_batch_dims:] ) ) , _out ) __a =1 for d in orig_batch_dims: flat_batch_dim *= d __a =flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(_snake_case : Optional[Any] ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t __a =0 __a =prepped_outputs for _ in range(__lowerCAmelCase ): # Chunk the input if not low_mem: __a =_select_chunk else: __a =partial( _chunk_slice , flat_start=__lowerCAmelCase , flat_end=min(__lowerCAmelCase , i + chunk_size ) , no_batch_dims=len(__lowerCAmelCase ) , ) __a =tensor_tree_map(__lowerCAmelCase , __lowerCAmelCase ) # Run the layer on the chunk __a =layer(**__lowerCAmelCase ) # Allocate space for the output if out is None: __a =tensor_tree_map(lambda _snake_case : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) , __lowerCAmelCase ) # Put the chunk in its pre-allocated space if isinstance(__lowerCAmelCase , __lowerCAmelCase ): def assign(_snake_case : Dict , _snake_case : Dict ) -> None: for k, v in da.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): assign(__lowerCAmelCase , da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: __a =da[k] assign(__lowerCAmelCase , __lowerCAmelCase ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): for xa, xa in zip(__lowerCAmelCase , __lowerCAmelCase ): if _add_into_out: xa[i : i + chunk_size] += xa else: __a =xa elif isinstance(__lowerCAmelCase , torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: __a =output_chunk else: raise ValueError('Not supported' ) i += chunk_size __a =tensor_tree_map(lambda _snake_case : t.view(orig_batch_dims + t.shape[1:] ) , __lowerCAmelCase ) return out class __magic_name__ : def __init__( self , __snake_case = 512 , ) -> Union[str, Any]: '''simple docstring''' __a =max_chunk_size __a =None __a =None def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> int: '''simple docstring''' logging.info('Tuning chunk size...' ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size __a =[2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] __a =[c for c in candidates if c > min_chunk_size] __a =[min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(__snake_case ) -> bool: try: with torch.no_grad(): fn(*__snake_case , chunk_size=__snake_case ) return True except RuntimeError: return False __a =0 __a =len(__snake_case ) - 1 while i > min_viable_chunk_size_index: __a =test_chunk_size(candidates[i] ) if not viable: __a =(min_viable_chunk_size_index + i) // 2 else: __a =i __a =(i + len(__snake_case ) - 1) // 2 return candidates[min_viable_chunk_size_index] def __magic_name__ ( self , __snake_case , __snake_case ) -> List[str]: '''simple docstring''' __a =True for aa, aa in zip(__snake_case , __snake_case ): assert type(__snake_case ) == type(__snake_case ) if isinstance(__snake_case , (list, tuple) ): consistent &= self._compare_arg_caches(__snake_case , __snake_case ) elif isinstance(__snake_case , __snake_case ): __a =[v for _, v in sorted(aa.items() , key=lambda __snake_case : x[0] )] __a =[v for _, v in sorted(aa.items() , key=lambda __snake_case : x[0] )] consistent &= self._compare_arg_caches(__snake_case , __snake_case ) else: consistent &= aa == aa return consistent def __magic_name__ ( self , __snake_case , __snake_case , __snake_case , ) -> List[Any]: '''simple docstring''' __a =True __a =tree_map(lambda __snake_case : a.shape if isinstance(__snake_case , torch.Tensor ) else a , __snake_case , __snake_case ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(__snake_case ) __a =self._compare_arg_caches(self.cached_arg_data , __snake_case ) else: # Otherwise, we can reuse the precomputed value __a =False if not consistent: __a =self._determine_favorable_chunk_size( __snake_case , __snake_case , __snake_case , ) __a =arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class _snake_case ( snake_case__ ): lowerCamelCase__: Tuple = 42 lowerCamelCase__: Dict = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.26.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version('''>=''', '''0.0.12''') ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class _snake_case ( snake_case__ ): lowerCamelCase__: Tuple = 42 lowerCamelCase__: List[Any] = 42 from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
157
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=3 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=10 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[1, 1, 2, 1] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase="relu" , UpperCAmelCase=3 , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = embeddings_size _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = scope _UpperCAmelCase = len(UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = TFResNetModel(config=UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFResNetForImageClassification(UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase = layer_type _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) @slow def UpperCamelCase ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TFResNetModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __A ( )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" @cached_property def UpperCamelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=UpperCAmelCase , return_tensors='tf' ) # forward pass _UpperCAmelCase = model(**UpperCAmelCase ) # verify the logits _UpperCAmelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _UpperCAmelCase = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCAmelCase , atol=1e-4 ) )
39
0
'''simple docstring''' def __a(SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Any ): '''simple docstring''' return base * power(__lowerCAmelCase , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print("Raise base to the power of exponent using recursion...") _SCREAMING_SNAKE_CASE = int(input("Enter the base: ").strip()) _SCREAMING_SNAKE_CASE = int(input("Enter the exponent: ").strip()) _SCREAMING_SNAKE_CASE = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents _SCREAMING_SNAKE_CASE = 1 / result print(f'''{base} to the power of {exponent} is {result}''')
158
def __A ( __lowerCAmelCase )-> list: """simple docstring""" if len(__lowerCAmelCase ) < 2: return collection def circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , mid + 1 , __lowerCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) - 1 ) return collection if __name__ == "__main__": _a = input('''Enter numbers separated by a comma:\n''').strip() _a = [int(item) for item in user_input.split(''',''')] print(circle_sort(unsorted))
39
0
from sklearn.metrics import recall_score import datasets lowerCamelCase__ = """ Recall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation: Recall = TP / (TP + FN) Where TP is the true positives and FN is the false negatives. """ lowerCamelCase__ = """ Args: - **predictions** (`list` of `int`): The predicted labels. - **references** (`list` of `int`): The ground truth labels. - **labels** (`list` of `int`): The set of labels to include when `average` is not set to `binary`, and their order when average is `None`. Labels present in the data can be excluded in this input, for example to calculate a multiclass average ignoring a majority negative class, while labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in y_true and y_pred are used in sorted order. Defaults to None. - **pos_label** (`int`): The class label to use as the \'positive class\' when calculating the recall. Defaults to `1`. - **average** (`string`): This parameter is required for multiclass/multilabel targets. If None, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`. - `\'binary\'`: Only report results for the class specified by `pos_label`. This is applicable only if the target labels and predictions are binary. - `\'micro\'`: Calculate metrics globally by counting the total true positives, false negatives, and false positives. - `\'macro\'`: Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - `\'weighted\'`: Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. Note that it can result in an F-score that is not between precision and recall. - `\'samples\'`: Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). - **sample_weight** (`list` of `float`): Sample weights Defaults to `None`. - **zero_division** (): Sets the value to return when there is a zero division. Defaults to . - `\'warn\'`: If there is a zero division, the return value is `0`, but warnings are also raised. - `0`: If there is a zero division, the return value is `0`. - `1`: If there is a zero division, the return value is `1`. Returns: - **recall** (`float`, or `array` of `float`): Either the general recall score, or the recall scores for individual classes, depending on the values input to `labels` and `average`. Minimum possible value is 0. Maximum possible value is 1. A higher recall means that more of the positive examples have been labeled correctly. Therefore, a higher recall is generally considered better. Examples: Example 1-A simple example with some errors >>> recall_metric = datasets.load_metric(\'recall\') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1]) >>> print(results) {\'recall\': 0.6666666666666666} Example 2-The same example as Example 1, but with `pos_label=0` instead of the default `pos_label=1`. >>> recall_metric = datasets.load_metric(\'recall\') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], pos_label=0) >>> print(results) {\'recall\': 0.5} Example 3-The same example as Example 1, but with `sample_weight` included. >>> recall_metric = datasets.load_metric(\'recall\') >>> sample_weight = [0.9, 0.2, 0.9, 0.3, 0.8] >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], sample_weight=sample_weight) >>> print(results) {\'recall\': 0.55} Example 4-A multiclass example, using different averages. >>> recall_metric = datasets.load_metric(\'recall\') >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = recall_metric.compute(predictions=predictions, references=references, average=\'macro\') >>> print(results) {\'recall\': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=\'micro\') >>> print(results) {\'recall\': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=\'weighted\') >>> print(results) {\'recall\': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {\'recall\': array([1., 0., 0.])} """ lowerCamelCase__ = """ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def _lowerCamelCase ( self : Any ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('int32' ) ), 'references': datasets.Sequence(datasets.Value('int32' ) ), } if self.config_name == 'multilabel' else { 'predictions': datasets.Value('int32' ), 'references': datasets.Value('int32' ), } ) , reference_urls=['https://scikit-learn.org/stable/modules/generated/sklearn.metrics.recall_score.html'] , ) def _lowerCamelCase ( self : Any , a : Tuple , a : Union[str, Any] , a : Optional[int]=None , a : List[str]=1 , a : List[Any]="binary" , a : List[Any]=None , a : str="warn" , ): '''simple docstring''' lowerCAmelCase__ : str = recall_score( a , a , labels=a , pos_label=a , average=a , sample_weight=a , zero_division=a , ) return {"recall": float(a ) if score.size == 1 else score}
212
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "Pix2StructImageProcessor" UpperCamelCase__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = False super().__init__(UpperCAmelCase , UpperCAmelCase ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 2048 , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , **UpperCAmelCase ) else: # add pixel_values and bbox _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , header_text=UpperCAmelCase , **UpperCAmelCase ) if text is not None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) if "attention_mask" in text_encoding: _UpperCAmelCase = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: _UpperCAmelCase = text_encoding.pop('input_ids' ) else: _UpperCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase ) return encoding_image_processor def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
from __future__ import annotations from collections.abc import Iterator from typing import Any class UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] ,A : Union[str, Any] ): __A = data __A = None class UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] ): __A = None __A = None def __iter__( self : Dict ): __A = self.head while self.head: yield node.data __A = node.next if node == self.head: break def __len__( self : Union[str, Any] ): return sum(1 for _ in self ) def __repr__( self : Optional[int] ): return "->".join(str(A ) for item in iter(self ) ) def UpperCamelCase_ ( self : Dict ,A : Union[str, Any] ): self.insert_nth(len(self ) ,A ) def UpperCamelCase_ ( self : Tuple ,A : Optional[int] ): self.insert_nth(0 ,A ) def UpperCamelCase_ ( self : str ,A : Dict ,A : List[Any] ): if index < 0 or index > len(self ): raise IndexError("list index out of range." ) __A = Node(A ) if self.head is None: __A = new_node # first node points itself __A = __A = new_node elif index == 0: # insert at head __A = self.head __A = __A = new_node else: __A = self.head for _ in range(index - 1 ): __A = temp.next __A = temp.next __A = new_node if index == len(self ) - 1: # insert at tail __A = new_node def UpperCamelCase_ ( self : Optional[Any] ): return self.delete_nth(0 ) def UpperCamelCase_ ( self : Any ): return self.delete_nth(len(self ) - 1 ) def UpperCamelCase_ ( self : Optional[int] ,A : List[str] = 0 ): if not 0 <= index < len(self ): raise IndexError("list index out of range." ) __A = self.head if self.head == self.tail: # just one node __A = __A = None elif index == 0: # delete head node __A = self.tail.next.next __A = self.head.next else: __A = self.head for _ in range(index - 1 ): __A = temp.next __A = temp.next __A = temp.next.next if index == len(self ) - 1: # delete at tail __A = temp return delete_node.data def UpperCamelCase_ ( self : List[Any] ): return len(self ) == 0 def UpperCAmelCase ( ) -> None: """simple docstring""" __A = CircularLinkedList() assert len(__lowerCAmelCase ) == 0 assert circular_linked_list.is_empty() is True assert str(__lowerCAmelCase ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(__lowerCAmelCase ) == i circular_linked_list.insert_nth(__lowerCAmelCase , i + 1 ) assert str(__lowerCAmelCase ) == "->".join(str(__lowerCAmelCase ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(__lowerCAmelCase ) == "->".join(str(__lowerCAmelCase ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(__lowerCAmelCase ) == "->".join(str(__lowerCAmelCase ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(__lowerCAmelCase ) == "->".join(str(__lowerCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(__lowerCAmelCase ) == "->".join(str(__lowerCAmelCase ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
15
class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase = "" , UpperCAmelCase = False ): """simple docstring""" _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 0 for q, w in zip(self.prefix , UpperCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=UpperCAmelCase , is_leaf=UpperCAmelCase ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def UpperCamelCase ( self , UpperCAmelCase = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = RadixNode() root.insert_many(__lowerCAmelCase ) assert all(root.find(__lowerCAmelCase ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = 'banana bananas bandanas bandana band apple all beast'.split() root.insert_many(__lowerCAmelCase ) print('Words:' , __lowerCAmelCase ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
39
0
import os import sys __lowerCamelCase : Union[str, Any] = os.path.join(os.path.dirname(__file__), '''src''') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __lowerCamelCase : Optional[int] = [ '''torch''', '''numpy''', '''tokenizers''', '''filelock''', '''requests''', '''tqdm''', '''regex''', '''sentencepiece''', '''sacremoses''', '''importlib_metadata''', '''huggingface_hub''', ] @add_start_docstrings(AutoConfig.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : List[Any] , **__UpperCamelCase : Any ) -> str: """simple docstring""" return AutoConfig.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : List[str] , **__UpperCamelCase : Tuple ) -> Optional[int]: """simple docstring""" return AutoTokenizer.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : Dict , **__UpperCamelCase : List[str] ) -> int: """simple docstring""" return AutoModel.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : Optional[Any] , **__UpperCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" return AutoModelForCausalLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : str , **__UpperCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : Dict , **__UpperCamelCase : str ) -> Optional[Any]: """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def __SCREAMING_SNAKE_CASE ( *__UpperCamelCase : List[Any] , **__UpperCamelCase : List[Any] ) -> Tuple: """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase )
219
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class __lowerCamelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments UpperCAmelCase="<s>" , UpperCAmelCase="<pad>" , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = bos, unk, pad, eos _UpperCAmelCase = [] _UpperCAmelCase = [] _UpperCAmelCase = {} _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , UpperCAmelCase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , UpperCAmelCase ): """simple docstring""" return sym in self.indices @classmethod def UpperCamelCase ( cls , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = cls() d.add_from_file(UpperCAmelCase ) return d def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase=1 , UpperCAmelCase=False ): """simple docstring""" if word in self.indices and not overwrite: _UpperCAmelCase = self.indices[word] _UpperCAmelCase = self.count[idx] + n return idx else: _UpperCAmelCase = len(self.symbols ) _UpperCAmelCase = idx self.symbols.append(UpperCAmelCase ) self.count.append(UpperCAmelCase ) return idx def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return 0 def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(UpperCAmelCase ) ) return _UpperCAmelCase = f.readlines() _UpperCAmelCase = self._load_meta(UpperCAmelCase ) for line in lines[indices_start_line:]: try: _UpperCAmelCase , _UpperCAmelCase = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": _UpperCAmelCase = True _UpperCAmelCase , _UpperCAmelCase = line.rsplit(' ' , 1 ) else: _UpperCAmelCase = False _UpperCAmelCase = int(UpperCAmelCase ) _UpperCAmelCase = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(UpperCAmelCase ) ) self.add_symbol(UpperCAmelCase , n=UpperCAmelCase , overwrite=UpperCAmelCase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" _UpperCAmelCase = dict((re.sub(R'@@$' , '' , __lowerCAmelCase ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , __lowerCAmelCase ), v) for k, v in d.items() ) _UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] _UpperCAmelCase = d[k] # restore return da def __A ( __lowerCAmelCase , __lowerCAmelCase )-> str: """simple docstring""" if not os.path.exists(__lowerCAmelCase ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'checkpoint.pt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location='cpu' ) _UpperCAmelCase = chkpt['cfg']['model'] # dicts _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'dict.txt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) _UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) _UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) _UpperCAmelCase = len(__lowerCAmelCase ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['vocab_file'] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # merges_file (bpecodes) _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'bpecodes' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowerCAmelCase , __lowerCAmelCase ) # model config _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'config.json' ) _UpperCAmelCase = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1E-12, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # tokenizer config _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 1_024, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # model _UpperCAmelCase = chkpt['model'] # remove unneeded keys _UpperCAmelCase = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) else: _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase = BioGptConfig.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase = BioGptForCausalLM(__lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(__lowerCAmelCase ) # save _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) print('Conversion is done!' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
39
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Tuple ) -> bool: '''simple docstring''' A__ = [int(__lowerCAmelCase ) for i in ip_va_address.split("." ) if i.isdigit()] return len(__lowerCAmelCase ) == 4 and all(0 <= int(__lowerCAmelCase ) <= 2_5_4 for octet in octets ) if __name__ == "__main__": lowerCAmelCase__ = input().strip() lowerCAmelCase__ = """valid""" if is_ip_va_address_valid(ip) else """invalid""" print(f"""{ip} is a {valid_or_invalid} IP v4 address.""")
68
from __future__ import annotations import collections import pprint from pathlib import Path def __A ( __lowerCAmelCase )-> str: """simple docstring""" return "".join(sorted(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase )-> list[str]: """simple docstring""" return word_by_signature[signature(__lowerCAmelCase )] _a = Path(__file__).parent.joinpath('''words.txt''').read_text(encoding='''utf-8''') _a = sorted({word.strip().lower() for word in data.splitlines()}) _a = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _a = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('''anagrams.txt''', '''w''') as file: file.write('''all_anagrams = \n ''') file.write(pprint.pformat(all_anagrams))
39
0
def lowerCamelCase__ ( a ) -> int: if not grid or not grid[0]: raise TypeError('''The grid does not contain the appropriate information''' ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _A: List[Any] = grid[0] for row_n in range(1 , len(__lowerCAmelCase ) ): _A: str = grid[row_n] _A: Dict = fill_row(__lowerCAmelCase , __lowerCAmelCase ) _A: List[str] = grid[row_n] return grid[-1][-1] def lowerCamelCase__ ( a , a ) -> list: current_row[0] += row_above[0] for cell_n in range(1 , len(__lowerCAmelCase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
121
from __future__ import annotations def __A ( __lowerCAmelCase )-> list[int]: """simple docstring""" _UpperCAmelCase = 2 _UpperCAmelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCAmelCase ) if n > 1: factors.append(__lowerCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
39
0
"""simple docstring""" from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING _a = logging.get_logger(__name__) @add_end_docstrings(snake_case__ ) class _UpperCAmelCase( snake_case__ ): def __init__( self , **__a) -> List[Any]: '''simple docstring''' super().__init__(**__a) requires_backends(self , '''vision''') requires_backends(self , '''torch''') if self.framework != "pt": raise ValueError(F'''The {self.__class__} is only available in PyTorch.''') self.check_model_type(__a) def UpperCAmelCase ( self , **__a) -> Tuple: '''simple docstring''' _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} # preprocess args if "points_per_batch" in kwargs: _UpperCamelCase = kwargs['''points_per_batch'''] if "points_per_crop" in kwargs: _UpperCamelCase = kwargs['''points_per_crop'''] if "crops_n_layers" in kwargs: _UpperCamelCase = kwargs['''crops_n_layers'''] if "crop_overlap_ratio" in kwargs: _UpperCamelCase = kwargs['''crop_overlap_ratio'''] if "crop_n_points_downscale_factor" in kwargs: _UpperCamelCase = kwargs['''crop_n_points_downscale_factor'''] # postprocess args if "pred_iou_thresh" in kwargs: _UpperCamelCase = kwargs['''pred_iou_thresh'''] if "stability_score_offset" in kwargs: _UpperCamelCase = kwargs['''stability_score_offset'''] if "mask_threshold" in kwargs: _UpperCamelCase = kwargs['''mask_threshold'''] if "stability_score_thresh" in kwargs: _UpperCamelCase = kwargs['''stability_score_thresh'''] if "crops_nms_thresh" in kwargs: _UpperCamelCase = kwargs['''crops_nms_thresh'''] if "output_rle_mask" in kwargs: _UpperCamelCase = kwargs['''output_rle_mask'''] if "output_bboxes_mask" in kwargs: _UpperCamelCase = kwargs['''output_bboxes_mask'''] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self , __a , *__a , __a=None , __a=None , **__a) -> int: '''simple docstring''' return super().__call__(__a , *__a , num_workers=__a , batch_size=__a , **__a) def UpperCAmelCase ( self , __a , __a=64 , __a = 0 , __a = 5_12 / 15_00 , __a = 32 , __a = 1 , ) -> List[Any]: '''simple docstring''' _UpperCamelCase = load_image(__a) _UpperCamelCase = self.image_processor.size['''longest_edge'''] _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.image_processor.generate_crop_boxes( __a , __a , __a , __a , __a , __a) _UpperCamelCase = self.image_processor(images=__a , return_tensors='''pt''') with self.device_placement(): if self.framework == "pt": _UpperCamelCase = self.get_inference_context() with inference_context(): _UpperCamelCase = self._ensure_tensor_on_device(__a , device=self.device) _UpperCamelCase = self.model.get_image_embeddings(model_inputs.pop('''pixel_values''')) _UpperCamelCase = image_embeddings _UpperCamelCase = grid_points.shape[1] _UpperCamelCase = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( '''Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. ''' '''To return all points at once, set points_per_batch to None''') for i in range(0 , __a , __a): _UpperCamelCase = grid_points[:, i : i + points_per_batch, :, :] _UpperCamelCase = input_labels[:, i : i + points_per_batch] _UpperCamelCase = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCAmelCase ( self , __a , __a=0.88 , __a=0.95 , __a=0 , __a=1 , ) -> str: '''simple docstring''' _UpperCamelCase = model_inputs.pop('''input_boxes''') _UpperCamelCase = model_inputs.pop('''is_last''') _UpperCamelCase = model_inputs.pop('''original_sizes''').tolist() _UpperCamelCase = model_inputs.pop('''reshaped_input_sizes''').tolist() _UpperCamelCase = self.model(**__a) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _UpperCamelCase = model_outputs['''pred_masks'''] _UpperCamelCase = self.image_processor.post_process_masks( __a , __a , __a , __a , binarize=__a) _UpperCamelCase = model_outputs['''iou_scores'''] _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __a , __a , __a , __a , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCAmelCase ( self , __a , __a=False , __a=False , __a=0.7 , ) -> List[Any]: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = [] for model_output in model_outputs: all_scores.append(model_output.pop('''iou_scores''')) all_masks.extend(model_output.pop('''masks''')) all_boxes.append(model_output.pop('''boxes''')) _UpperCamelCase = torch.cat(__a) _UpperCamelCase = torch.cat(__a) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.image_processor.post_process_for_mask_generation( __a , __a , __a , __a) _UpperCamelCase = defaultdict(__a) for output in model_outputs: for k, v in output.items(): extra[k].append(__a) _UpperCamelCase = {} if output_rle_mask: _UpperCamelCase = rle_mask if output_bboxes_mask: _UpperCamelCase = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
194
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __A ( )-> tuple[list[int], int]: """simple docstring""" _UpperCAmelCase = [randint(-1_000 , 1_000 ) for i in range(10 )] _UpperCAmelCase = randint(-5_000 , 5_000 ) return (arr, r) _a = make_dataset() def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, ...]: """simple docstring""" for triplet in permutations(__lowerCAmelCase , 3 ): if sum(__lowerCAmelCase ) == target: return tuple(sorted(__lowerCAmelCase ) ) return (0, 0, 0) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, int, int]: """simple docstring""" arr.sort() _UpperCAmelCase = len(__lowerCAmelCase ) for i in range(n - 1 ): _UpperCAmelCase , _UpperCAmelCase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __A ( )-> tuple[float, float]: """simple docstring""" _UpperCAmelCase = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' _UpperCAmelCase = '\ntriplet_sum1(*dataset)\n' _UpperCAmelCase = '\ntriplet_sum2(*dataset)\n' _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) return (min(__lowerCAmelCase ), min(__lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _a = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
39
0
def lowerCamelCase_ ( _UpperCamelCase ) -> str: """simple docstring""" snake_case_ : Optional[Any] = '''''' for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def lowerCamelCase_ ( _UpperCamelCase ) -> dict[str, str]: """simple docstring""" snake_case_ : Optional[Any] = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key snake_case_ : List[Any] = remove_duplicates(key.upper() ) snake_case_ : Optional[Any] = len(__lowerCAmelCase ) # First fill cipher with key characters snake_case_ : Optional[int] = {alphabet[i]: char for i, char in enumerate(__lowerCAmelCase )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(__lowerCAmelCase ) , 26 ): snake_case_ : Optional[int] = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 snake_case_ : Any = alphabet[i - offset] snake_case_ : List[Any] = char return cipher_alphabet def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> str: """simple docstring""" return "".join(cipher_map.get(__lowerCAmelCase , __lowerCAmelCase ) for ch in message.upper() ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> str: """simple docstring""" snake_case_ : Union[str, Any] = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(__lowerCAmelCase , __lowerCAmelCase ) for ch in message.upper() ) def lowerCamelCase_ ( ) -> None: """simple docstring""" snake_case_ : Dict = input('''Enter message to encode or decode: ''' ).strip() snake_case_ : str = input('''Enter keyword: ''' ).strip() snake_case_ : List[Any] = input('''Encipher or decipher? E/D:''' ).strip()[0].lower() try: snake_case_ : Union[str, Any] = {'''e''': encipher, '''d''': decipher}[option] except KeyError: raise KeyError('''invalid input option''' ) snake_case_ : Union[str, Any] = create_cipher_map(__lowerCAmelCase ) print(func(__lowerCAmelCase , __lowerCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
279
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase , cache_dir=UpperCAmelCase ) _UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(UpperCAmelCase , os.listdir(UpperCAmelCase )[0] , 'snapshots' ) )] _UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 4 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1e-3 assert np.abs(np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5e-1 _UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(UpperCAmelCase ) == num_samples def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , set_alpha_to_one=UpperCAmelCase , steps_offset=1 , ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=UpperCAmelCase , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = scheduler.create_state() _UpperCAmelCase = scheduler_state _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # With memory efficient attention _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , use_memory_efficient_attention=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
39
0
import os import pytest from attr import dataclass lowercase__ ='us-east-1' # defaults region @dataclass class UpperCamelCase__ : _SCREAMING_SNAKE_CASE : str = 42 _SCREAMING_SNAKE_CASE : List[str] = "arn:aws:iam::558105141721:role/sagemaker_execution_role" _SCREAMING_SNAKE_CASE : Tuple = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 500, "save_steps": 5_500, } _SCREAMING_SNAKE_CASE : Any = {**hyperparameters, "max_steps": 1_000} @property def lowerCAmelCase (self : Optional[Any] ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def lowerCAmelCase (self : List[str] ): return f"{self.framework}-transfromers-test" @property def lowerCAmelCase (self : List[Any] ): return f"./tests/sagemaker/scripts/{self.framework}" @property def lowerCAmelCase (self : int ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] ): __a : int = SageMakerTestEnvironment(framework=request.cls.framework )
216
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = AlbertTokenizer UpperCamelCase__ = AlbertTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = True def UpperCamelCase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 'this is a test' _UpperCAmelCase = 'this is a test' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = '<pad>' _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(UpperCAmelCase ) , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = 'I was born in 92000, and this is falsé.' _UpperCAmelCase = tokenizer.tokenize(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [48, 25, 21, 1289] ) _UpperCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) _UpperCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode('sequence builders' ) _UpperCAmelCase = tokenizer.encode('multi-sequence build' ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
39
0
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __magic_name__ ( snake_case__ ): def __magic_name__ ( self ) -> str: '''simple docstring''' __a =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__snake_case , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__snake_case , 'num_attention_heads' ) ) self.parent.assertTrue(hasattr(__snake_case , 'num_encoder_blocks' ) ) class __magic_name__ : def __init__( self , __snake_case , __snake_case=13 , __snake_case=64 , __snake_case=3 , __snake_case=4 , __snake_case=[2, 2, 2, 2] , __snake_case=[8, 4, 2, 1] , __snake_case=[16, 32, 64, 128] , __snake_case=[1, 4, 8, 16] , __snake_case=[1, 2, 4, 8] , __snake_case=True , __snake_case=True , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=0.02 , __snake_case=3 , __snake_case=None , ) -> Tuple: '''simple docstring''' __a =parent __a =batch_size __a =image_size __a =num_channels __a =num_encoder_blocks __a =sr_ratios __a =depths __a =hidden_sizes __a =downsampling_rates __a =num_attention_heads __a =is_training __a =use_labels __a =hidden_act __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =initializer_range __a =num_labels __a =scope def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a =None if self.use_labels: __a =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __a =self.get_config() return config, pixel_values, labels def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> List[str]: '''simple docstring''' __a =SegformerModel(config=__snake_case ) model.to(__snake_case ) model.eval() __a =model(__snake_case ) __a =__a =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> int: '''simple docstring''' __a =self.num_labels __a =SegformerForSemanticSegmentation(__snake_case ) model.to(__snake_case ) model.eval() __a =model(__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) __a =model(__snake_case , labels=__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case ) -> str: '''simple docstring''' __a =1 __a =SegformerForSemanticSegmentation(config=__snake_case ) model.to(__snake_case ) model.eval() __a =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(__snake_case ) __a =model(__snake_case , labels=__snake_case ) self.parent.assertGreater(result.loss , 0.0 ) def __magic_name__ ( self ) -> Tuple: '''simple docstring''' __a =self.prepare_config_and_inputs() __a , __a , __a =config_and_inputs __a ={'pixel_values': pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( snake_case__ , snake_case__ , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SegformerModel, 'image-classification': SegformerForImageClassification, 'image-segmentation': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def __magic_name__ ( self ) -> int: '''simple docstring''' __a =SegformerModelTester(self ) __a =SegformerConfigTester(self , config_class=__snake_case ) def __magic_name__ ( self ) -> str: '''simple docstring''' self.config_tester.run_common_tests() def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*__snake_case ) def __magic_name__ ( self ) -> List[str]: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*__snake_case ) @unittest.skip('SegFormer does not use inputs_embeds' ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' pass @unittest.skip('SegFormer does not have get_input_embeddings method and get_output_embeddings methods' ) def __magic_name__ ( self ) -> Dict: '''simple docstring''' pass def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a , __a =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a =model_class(__snake_case ) __a =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a =[*signature.parameters.keys()] __a =['pixel_values'] self.assertListEqual(arg_names[:1] , __snake_case ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' __a , __a =self.model_tester.prepare_config_and_inputs_for_common() __a =True for model_class in self.all_model_classes: __a =True __a =False __a =True __a =model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): __a =model(**self._prepare_for_class(__snake_case , __snake_case ) ) __a =outputs.attentions __a =sum(self.model_tester.depths ) self.assertEqual(len(__snake_case ) , __snake_case ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a =True __a =model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): __a =model(**self._prepare_for_class(__snake_case , __snake_case ) ) __a =outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) __a =(self.model_tester.image_size // 4) ** 2 __a =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) __a =(self.model_tester.image_size // 32) ** 2 __a =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) __a =len(__snake_case ) # Check attention is always last and order is fine __a =True __a =True __a =model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): __a =model(**self._prepare_for_class(__snake_case , __snake_case ) ) self.assertEqual(out_len + 1 , len(__snake_case ) ) __a =outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) __a =(self.model_tester.image_size // 4) ** 2 __a =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def __magic_name__ ( self ) -> Any: '''simple docstring''' def check_hidden_states_output(__snake_case , __snake_case , __snake_case ): __a =model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): __a =model(**self._prepare_for_class(__snake_case , __snake_case ) ) __a =outputs.hidden_states __a =self.model_tester.num_encoder_blocks self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __a , __a =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a =True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a =True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) def __magic_name__ ( self ) -> List[str]: '''simple docstring''' if not self.model_tester.is_training: return __a , __a =self.model_tester.prepare_config_and_inputs_for_common() __a =True for model_class in self.all_model_classes: if model_class in get_values(__snake_case ): continue __a =model_class(__snake_case ) model.to(__snake_case ) model.train() __a =self._prepare_for_class(__snake_case , __snake_case , return_labels=__snake_case ) __a =model(**__snake_case ).loss loss.backward() @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __magic_name__ ( self ) -> List[str]: '''simple docstring''' pass @slow def __magic_name__ ( self ) -> Tuple: '''simple docstring''' for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a =SegformerModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def UpperCamelCase_( ): """simple docstring""" __a =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class __magic_name__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) __a =SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( __snake_case ) __a =prepare_img() __a =image_processor(images=__snake_case , return_tensors='pt' ) __a =encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): __a =model(__snake_case ) __a =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , __snake_case ) __a =torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1e-4 ) ) @slow def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) __a =SegformerForSemanticSegmentation.from_pretrained( 'nvidia/segformer-b1-finetuned-cityscapes-1024-1024' ).to(__snake_case ) __a =prepare_img() __a =image_processor(images=__snake_case , return_tensors='pt' ) __a =encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): __a =model(__snake_case ) __a =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , __snake_case ) __a =torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1e-1 ) ) @slow def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' __a =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) __a =SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( __snake_case ) __a =prepare_img() __a =image_processor(images=__snake_case , return_tensors='pt' ) __a =encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): __a =model(__snake_case ) __a =outputs.logits.detach().cpu() __a =image_processor.post_process_semantic_segmentation(outputs=__snake_case , target_sizes=[(500, 300)] ) __a =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , __snake_case ) __a =image_processor.post_process_semantic_segmentation(outputs=__snake_case ) __a =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , __snake_case )
218
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
import copy import random from transformers import CLIPTokenizer class _snake_case ( snake_case__ ): def __init__( self: List[str] , *__lowerCamelCase: Optional[int] , **__lowerCamelCase: int ) -> List[str]: super().__init__(*__lowerCamelCase , **__lowerCamelCase ) __UpperCAmelCase : List[Any] = {} def _lowerCamelCase ( self: str , __lowerCamelCase: List[str] , *__lowerCamelCase: Dict , **__lowerCamelCase: List[str] ) -> Optional[Any]: __UpperCAmelCase : Optional[Any] = super().add_tokens(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) if num_added_tokens == 0: raise ValueError( f'''The tokenizer already contains the token {placeholder_token}. Please pass a different''' " `placeholder_token` that is not already in the tokenizer." ) def _lowerCamelCase ( self: Dict , __lowerCamelCase: int , *__lowerCamelCase: int , __lowerCamelCase: Dict=1 , **__lowerCamelCase: Any ) -> List[Any]: __UpperCAmelCase : Tuple = [] if num_vec_per_token == 1: self.try_adding_tokens(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) output.append(__lowerCamelCase ) else: __UpperCAmelCase : Tuple = [] for i in range(__lowerCamelCase ): __UpperCAmelCase : Union[str, Any] = placeholder_token + f'''_{i}''' self.try_adding_tokens(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) output.append(__lowerCamelCase ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f'''The tokenizer already has placeholder token {token} that can get confused with''' f''' {placeholder_token}keep placeholder tokens independent''' ) __UpperCAmelCase : List[Any] = output def _lowerCamelCase ( self: Tuple , __lowerCamelCase: int , __lowerCamelCase: Optional[int]=False , __lowerCamelCase: Any=1.0 ) -> List[Any]: if isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[Any] = [] for i in range(len(__lowerCamelCase ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=__lowerCamelCase ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: __UpperCAmelCase : str = self.token_map[placeholder_token] __UpperCAmelCase : Dict = tokens[: 1 + int(len(__lowerCamelCase ) * prop_tokens_to_load )] if vector_shuffle: __UpperCAmelCase : Any = copy.copy(__lowerCamelCase ) random.shuffle(__lowerCamelCase ) __UpperCAmelCase : int = text.replace(__lowerCamelCase , " ".join(__lowerCamelCase ) ) return text def __call__( self: List[str] , __lowerCamelCase: Union[str, Any] , *__lowerCamelCase: int , __lowerCamelCase: Optional[Any]=False , __lowerCamelCase: Any=1.0 , **__lowerCamelCase: Union[str, Any] ) -> Union[str, Any]: return super().__call__( self.replace_placeholder_tokens_in_text( __lowerCamelCase , vector_shuffle=__lowerCamelCase , prop_tokens_to_load=__lowerCamelCase ) , *__lowerCamelCase , **__lowerCamelCase , ) def _lowerCamelCase ( self: Dict , __lowerCamelCase: Union[str, Any] , *__lowerCamelCase: int , __lowerCamelCase: Optional[int]=False , __lowerCamelCase: List[Any]=1.0 , **__lowerCamelCase: Any ) -> str: return super().encode( self.replace_placeholder_tokens_in_text( __lowerCamelCase , vector_shuffle=__lowerCamelCase , prop_tokens_to_load=__lowerCamelCase ) , *__lowerCamelCase , **__lowerCamelCase , )
157
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "distilbert" UpperCamelCase__ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , UpperCAmelCase=3_0522 , UpperCAmelCase=512 , UpperCAmelCase=False , UpperCAmelCase=6 , UpperCAmelCase=12 , UpperCAmelCase=768 , UpperCAmelCase=4 * 768 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=0.02 , UpperCAmelCase=0.1 , UpperCAmelCase=0.2 , UpperCAmelCase=0 , **UpperCAmelCase , ): """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = sinusoidal_pos_embds _UpperCAmelCase = n_layers _UpperCAmelCase = n_heads _UpperCAmelCase = dim _UpperCAmelCase = hidden_dim _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation _UpperCAmelCase = initializer_range _UpperCAmelCase = qa_dropout _UpperCAmelCase = seq_classif_dropout super().__init__(**UpperCAmelCase , pad_token_id=UpperCAmelCase ) class __lowerCamelCase ( snake_case__): """simple docstring""" @property def UpperCamelCase ( self ): """simple docstring""" if self.task == "multiple-choice": _UpperCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _UpperCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
39
0
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class lowerCAmelCase_ ( snake_case__ ): __lowerCamelCase : str = 42 __lowerCamelCase : Optional[Any] = 42 class lowerCAmelCase_ ( snake_case__ ,snake_case__ ): __lowerCamelCase : Optional[int] = 1 @register_to_config def __init__( self , _lowerCAmelCase = 2000 , _lowerCAmelCase = 0.15 , _lowerCAmelCase = 0.01 , _lowerCAmelCase = 1348.0 , _lowerCAmelCase = 1E-5 , _lowerCAmelCase = 1 , ) -> str: _lowerCAmelCase = sigma_max # setable values _lowerCAmelCase = None self.set_sigmas(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = None ) -> Union[str, Any]: return sample def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None ) -> str: _lowerCAmelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCAmelCase = torch.linspace(1 , _lowerCAmelCase , _lowerCAmelCase , device=_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None ) -> Optional[int]: _lowerCAmelCase = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCAmelCase = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCAmelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(_lowerCAmelCase , _lowerCAmelCase ) _lowerCAmelCase = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCAmelCase = torch.exp(torch.linspace(math.log(_lowerCAmelCase ) , math.log(_lowerCAmelCase ) , _lowerCAmelCase ) ) _lowerCAmelCase = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = True , ) -> Any: if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler" ) _lowerCAmelCase = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCAmelCase = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCAmelCase = timesteps.to(self.discrete_sigmas.device ) _lowerCAmelCase = self.discrete_sigmas[timesteps].to(sample.device ) _lowerCAmelCase = self.get_adjacent_sigma(_lowerCAmelCase , _lowerCAmelCase ).to(sample.device ) _lowerCAmelCase = torch.zeros_like(_lowerCAmelCase ) _lowerCAmelCase = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCAmelCase = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _lowerCAmelCase = diffusion.unsqueeze(-1 ) _lowerCAmelCase = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCAmelCase = randn_tensor( sample.shape , layout=sample.layout , generator=_lowerCAmelCase , device=sample.device , dtype=sample.dtype ) _lowerCAmelCase = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCAmelCase = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=_lowerCAmelCase , prev_sample_mean=_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = True , ) -> Dict: if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler" ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCAmelCase = randn_tensor(sample.shape , layout=sample.layout , generator=_lowerCAmelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _lowerCAmelCase = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _lowerCAmelCase = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _lowerCAmelCase = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCAmelCase = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCAmelCase = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _lowerCAmelCase = step_size.unsqueeze(-1 ) _lowerCAmelCase = sample + step_size * model_output _lowerCAmelCase = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> Dict: _lowerCAmelCase = timesteps.to(original_samples.device ) _lowerCAmelCase = self.discrete_sigmas.to(original_samples.device )[timesteps] _lowerCAmelCase = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(_lowerCAmelCase ) * sigmas[:, None, None, None] ) _lowerCAmelCase = noise + original_samples return noisy_samples def __len__( self ) -> Union[str, Any]: return self.config.num_train_timesteps
158
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() _a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCamelCase ( snake_case__): """simple docstring""" def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {'source': 'What is love ?', 'target': 'life'} _UpperCAmelCase = {'train': 12, 'val': 2, 'test': 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: _UpperCAmelCase = '\n'.join([contents[field]] * n_lines[split] ) with open(os.path.join(UpperCAmelCase , F"""{split}.{field}""" ) , 'w' ) as f: f.write(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase = "pytorch" ): """simple docstring""" _UpperCAmelCase = self.get_auto_remove_tmp_dir() _UpperCAmelCase = os.path.join(UpperCAmelCase , 'output' ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'data' ) self._create_dummy_data(data_dir=UpperCAmelCase ) _UpperCAmelCase = F""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(F"""--gpus={gpus}""" ) if is_apex_available(): testargs.append('--fp16' ) else: testargs.append('--gpus=0' ) testargs.append('--distributed_backend=ddp_cpu' ) testargs.append('--num_processes=2' ) _UpperCAmelCase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(UpperCAmelCase , env=self.get_env() ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'metrics.json' ) with open(UpperCAmelCase ) as f: _UpperCAmelCase = json.load(UpperCAmelCase ) return result @require_torch_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 )
39
0
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowerCamelCase__ = random.Random() if is_torch_available(): import torch def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=1.0 , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> Union[str, Any]: if rng is None: lowerCAmelCase__ : int = global_rng lowerCAmelCase__ : List[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class A__ ( unittest.TestCase ): def __init__( self : Dict , a : Optional[int] , a : Tuple=7 , a : Optional[Any]=400 , a : Optional[int]=2_000 , a : Any=1 , a : List[Any]=0.0 , a : Tuple=16_000 , a : Any=True , a : List[Any]=True , ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = parent lowerCAmelCase__ : Dict = batch_size lowerCAmelCase__ : int = min_seq_length lowerCAmelCase__ : str = max_seq_length lowerCAmelCase__ : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase__ : str = feature_size lowerCAmelCase__ : Optional[Any] = padding_value lowerCAmelCase__ : Union[str, Any] = sampling_rate lowerCAmelCase__ : List[Any] = return_attention_mask lowerCAmelCase__ : str = do_normalize def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowerCamelCase ( self : Tuple , a : List[Any]=False , a : str=False ): '''simple docstring''' def _flatten(a : Tuple ): return list(itertools.chain(*a ) ) if equal_length: lowerCAmelCase__ : Optional[int] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCAmelCase__ : Optional[Any] = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase__ : List[str] = [np.asarray(a ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A__ ( snake_case__ , unittest.TestCase ): lowercase = ASTFeatureExtractor def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = ASTFeatureExtractionTester(self ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase__ : Union[str, Any] = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] lowerCAmelCase__ : Union[str, Any] = [np.asarray(a ) for speech_input in speech_inputs] # Test not batched input lowerCAmelCase__ : Tuple = feat_extract(speech_inputs[0] , return_tensors='np' ).input_values lowerCAmelCase__ : int = feat_extract(np_speech_inputs[0] , return_tensors='np' ).input_values self.assertTrue(np.allclose(a , a , atol=1E-3 ) ) # Test batched lowerCAmelCase__ : int = feat_extract(a , padding=a , return_tensors='np' ).input_values lowerCAmelCase__ : Optional[int] = feat_extract(a , padding=a , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(a , a ): self.assertTrue(np.allclose(a , a , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase__ : Dict = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCAmelCase__ : List[Any] = np.asarray(a ) lowerCAmelCase__ : Any = feat_extract(a , return_tensors='np' ).input_values lowerCAmelCase__ : Optional[int] = feat_extract(a , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(a , a ): self.assertTrue(np.allclose(a , a , atol=1E-3 ) ) @require_torch def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' import torch lowerCAmelCase__ : int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Optional[int] = np.random.rand(100 ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase__ : Tuple = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCAmelCase__ : Union[str, Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def _lowerCamelCase ( self : Dict , a : Any ): '''simple docstring''' from datasets import load_dataset lowerCAmelCase__ : Dict = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCAmelCase__ : Dict = ds.sort('id' ).select(range(a ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] @require_torch def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = torch.tensor( [-0.9_8_9_4, -1.2_7_7_6, -0.9_0_6_6, -1.2_7_7_6, -0.9_3_4_9, -1.2_6_0_9, -1.0_3_8_6, -1.2_7_7_6, -1.1_5_6_1, -1.2_7_7_6, -1.2_0_5_2, -1.2_7_2_3, -1.2_1_9_0, -1.2_1_3_2, -1.2_7_7_6, -1.1_1_3_3, -1.1_9_5_3, -1.1_3_4_3, -1.1_5_8_4, -1.2_2_0_3, -1.1_7_7_0, -1.2_4_7_4, -1.2_3_8_1, -1.1_9_3_6, -0.9_2_7_0, -0.8_3_1_7, -0.8_0_4_9, -0.7_7_0_6, -0.7_5_6_5, -0.7_8_6_9] ) # fmt: on lowerCAmelCase__ : Optional[int] = self._load_datasamples(1 ) lowerCAmelCase__ : List[str] = ASTFeatureExtractor() lowerCAmelCase__ : Tuple = feature_extractor(a , return_tensors='pt' ).input_values self.assertEquals(input_values.shape , (1, 1_024, 128) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , a , atol=1E-4 ) )
212
class __lowerCamelCase : """simple docstring""" def __init__( self ): """simple docstring""" _UpperCAmelCase = {} # Mapping from char to TrieNode _UpperCAmelCase = False def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: _UpperCAmelCase = TrieNode() _UpperCAmelCase = curr.nodes[char] _UpperCAmelCase = True def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: return False _UpperCAmelCase = curr.nodes[char] return curr.is_leaf def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" def _delete(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: if index == len(UpperCAmelCase ): # If word does not exist if not curr.is_leaf: return False _UpperCAmelCase = False return len(curr.nodes ) == 0 _UpperCAmelCase = word[index] _UpperCAmelCase = curr.nodes.get(UpperCAmelCase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted _UpperCAmelCase = _delete(UpperCAmelCase , UpperCAmelCase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCAmelCase , 0 ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" if node.is_leaf: print(__lowerCAmelCase , end=' ' ) for key, value in node.nodes.items(): print_words(__lowerCAmelCase , word + key ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = TrieNode() root.insert_many(__lowerCAmelCase ) # print_words(root, "") assert all(root.find(__lowerCAmelCase ) for word in words ) assert root.find('banana' ) assert not root.find('bandanas' ) assert not root.find('apps' ) assert root.find('apple' ) assert root.find('all' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" print(str(__lowerCAmelCase ) , 'works!' if passes else 'doesn\'t work :(' ) def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" print_results('Testing trie functionality' , test_trie() ) if __name__ == "__main__": main()
39
0
def UpperCAmelCase ( a_ ) -> str: """simple docstring""" if number > 0: raise ValueError("input must be a negative integer" ) __A = len(bin(__lowerCAmelCase )[3:] ) __A = bin(abs(__lowerCAmelCase ) - (1 << binary_number_length) )[3:] __A = ( ( "1" + "0" * (binary_number_length - len(__lowerCAmelCase )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
15
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _a = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class __lowerCamelCase ( unittest.TestCase): """simple docstring""" UpperCamelCase__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = ZeroShotClassificationPipeline( model=UpperCAmelCase , tokenizer=UpperCAmelCase , candidate_labels=['polics', 'health'] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # No kwarg _UpperCAmelCase = classifier('Who are you voting for in 2020?' , ['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics, public health' ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics', 'public health'] ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='This text is about {}' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # https://github.com/huggingface/transformers/issues/13846 _UpperCAmelCase = classifier(['I am happy'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(1 ) ] , ) _UpperCAmelCase = classifier(['I am happy', 'I am sad'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(2 ) ] , ) with self.assertRaises(UpperCAmelCase ): classifier('' , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier(UpperCAmelCase , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels='' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='Not formatting template' , ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template=UpperCAmelCase , ) self.run_entailment_id(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = zero_shot_classifier.model.config _UpperCAmelCase = config.labelaid _UpperCAmelCase = zero_shot_classifier.entailment_id _UpperCAmelCase = {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) _UpperCAmelCase = {'entailment': 0, 'neutral': 1, 'contradiction': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 0, 'NON-ENTAIL': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 2, 'NEUTRAL': 1, 'CONTR': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) _UpperCAmelCase = original_labelaid self.assertEqual(UpperCAmelCase , zero_shot_classifier.entailment_id ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( 'Who are you voting for in 2020?' * 100 , candidate_labels=['politics', 'public health', 'science'] ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='tf' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @slow @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='pt' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , ) @slow @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='tf' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , )
39
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Any = logging.get_logger(__name__) __lowerCamelCase : Tuple = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class __snake_case ( snake_case__ ): lowerCAmelCase_ = "roformer" def __init__( self : List[Any] , _lowercase : str=5_00_00 , _lowercase : Optional[int]=None , _lowercase : Dict=7_68 , _lowercase : Any=12 , _lowercase : List[Any]=12 , _lowercase : str=30_72 , _lowercase : Optional[int]="gelu" , _lowercase : Any=0.1 , _lowercase : str=0.1 , _lowercase : str=15_36 , _lowercase : Tuple=2 , _lowercase : Tuple=0.02 , _lowercase : List[str]=1E-12 , _lowercase : List[str]=0 , _lowercase : int=False , _lowercase : Dict=True , **_lowercase : int , ): """simple docstring""" super().__init__(pad_token_id=_lowercase , **_lowercase ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size if embedding_size is None else embedding_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = rotary_value SCREAMING_SNAKE_CASE__ = use_cache class __snake_case ( snake_case__ ): @property def __a ( self : Tuple ): """simple docstring""" if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """sequence"""} SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
219
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _a = get_logger(__name__) class __lowerCamelCase ( enum.Enum): """simple docstring""" UpperCamelCase__ = "all_checks" UpperCamelCase__ = "basic_checks" UpperCamelCase__ = "no_checks" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None )-> str: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase = ' for ' + verification_name if verification_name is not None else '' if len(__lowerCAmelCase ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__lowerCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__lowerCAmelCase ) ) logger.info('All the splits matched successfully.' ) def __A ( __lowerCAmelCase , __lowerCAmelCase = True )-> dict: """simple docstring""" if record_checksum: _UpperCAmelCase = shaaaa() with open(__lowerCAmelCase , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__lowerCAmelCase ) _UpperCAmelCase = m.hexdigest() else: _UpperCAmelCase = None return {"num_bytes": os.path.getsize(__lowerCAmelCase ), "checksum": checksum} def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
39
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCAmelCase__ = logging.get_logger(__name__) if is_vision_available(): import PIL class a__ ( snake_case__ ): """simple docstring""" __lowerCamelCase = ['pixel_values'] def __init__( self , lowercase = True , lowercase = None , lowercase = PILImageResampling.BICUBIC , lowercase = True , lowercase = None , lowercase = True , lowercase = 1 / 255 , lowercase = True , lowercase = None , lowercase = None , lowercase = True , **lowercase , ) -> Optional[Any]: '''simple docstring''' super().__init__(**lowercase ) A__ = size if size is not None else {"shortest_edge": 224} A__ = get_size_dict(lowercase , default_to_square=lowercase ) A__ = crop_size if crop_size is not None else {"height": 224, "width": 224} A__ = get_size_dict(lowercase , default_to_square=lowercase , param_name="crop_size" ) A__ = do_resize A__ = size A__ = resample A__ = do_center_crop A__ = crop_size A__ = do_rescale A__ = rescale_factor A__ = do_normalize A__ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN A__ = image_std if image_std is not None else OPENAI_CLIP_STD A__ = do_convert_rgb def UpperCamelCase ( self , lowercase , lowercase , lowercase = PILImageResampling.BICUBIC , lowercase = None , **lowercase , ) -> List[str]: '''simple docstring''' A__ = get_size_dict(lowercase , default_to_square=lowercase ) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) A__ = get_resize_output_image_size(lowercase , size=size["shortest_edge"] , default_to_square=lowercase ) return resize(lowercase , size=lowercase , resample=lowercase , data_format=lowercase , **lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase = None , **lowercase , ) -> Union[str, Any]: '''simple docstring''' A__ = get_size_dict(lowercase ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(lowercase , size=(size["height"], size["width"]) , data_format=lowercase , **lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase = None , **lowercase , ) -> str: '''simple docstring''' return rescale(lowercase , scale=lowercase , data_format=lowercase , **lowercase ) def UpperCamelCase ( self , lowercase , lowercase , lowercase , lowercase = None , **lowercase , ) -> Union[str, Any]: '''simple docstring''' return normalize(lowercase , mean=lowercase , std=lowercase , data_format=lowercase , **lowercase ) def UpperCamelCase ( self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = ChannelDimension.FIRST , **lowercase , ) -> str: '''simple docstring''' A__ = do_resize if do_resize is not None else self.do_resize A__ = size if size is not None else self.size A__ = get_size_dict(lowercase , param_name="size" , default_to_square=lowercase ) A__ = resample if resample is not None else self.resample A__ = do_center_crop if do_center_crop is not None else self.do_center_crop A__ = crop_size if crop_size is not None else self.crop_size A__ = get_size_dict(lowercase , param_name="crop_size" , default_to_square=lowercase ) A__ = do_rescale if do_rescale is not None else self.do_rescale A__ = rescale_factor if rescale_factor is not None else self.rescale_factor A__ = do_normalize if do_normalize is not None else self.do_normalize A__ = image_mean if image_mean is not None else self.image_mean A__ = image_std if image_std is not None else self.image_std A__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb A__ = make_list_of_images(lowercase ) if not valid_images(lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: A__ = [convert_to_rgb(lowercase ) for image in images] # All transformations expect numpy arrays. A__ = [to_numpy_array(lowercase ) for image in images] if do_resize: A__ = [self.resize(image=lowercase , size=lowercase , resample=lowercase ) for image in images] if do_center_crop: A__ = [self.center_crop(image=lowercase , size=lowercase ) for image in images] if do_rescale: A__ = [self.rescale(image=lowercase , scale=lowercase ) for image in images] if do_normalize: A__ = [self.normalize(image=lowercase , mean=lowercase , std=lowercase ) for image in images] A__ = [to_channel_dimension_format(lowercase , lowercase ) for image in images] A__ = {"pixel_values": images} return BatchFeature(data=lowercase , tensor_type=lowercase )
68
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=16 , UpperCAmelCase=[1, 2, 1] , UpperCAmelCase=[2, 2, 4] , UpperCAmelCase=2 , UpperCAmelCase=2.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=10 , UpperCAmelCase=8 , UpperCAmelCase=["stage1", "stage2", "stage3"] , UpperCAmelCase=[1, 2, 3] , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = embed_dim _UpperCAmelCase = depths _UpperCAmelCase = num_heads _UpperCAmelCase = window_size _UpperCAmelCase = mlp_ratio _UpperCAmelCase = qkv_bias _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = drop_path_rate _UpperCAmelCase = hidden_act _UpperCAmelCase = use_absolute_embeddings _UpperCAmelCase = patch_norm _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = initializer_range _UpperCAmelCase = is_training _UpperCAmelCase = scope _UpperCAmelCase = use_labels _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = encoder_stride _UpperCAmelCase = out_features _UpperCAmelCase = out_indices def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) _UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCAmelCase ): _UpperCAmelCase = ['stem'] _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCamelCase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) # Swin has a different seq_length _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = 3 _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCAmelCase ): _UpperCAmelCase = 0 return t def check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase={} ): with torch.no_grad(): _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ).to_tuple() def recursive_check(UpperCAmelCase , UpperCAmelCase ): if isinstance(UpperCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase , UpperCAmelCase ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCAmelCase ) , set_nan_tensor_to_zero(UpperCAmelCase ) , atol=1e-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}. Dict has""" F""" `nan`: {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}.""" ) , ) recursive_check(UpperCAmelCase , UpperCAmelCase ) for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) @require_torch class __lowerCamelCase ( unittest.TestCase , snake_case__): """simple docstring""" UpperCamelCase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCamelCase__ = MaskFormerSwinConfig def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: _UpperCAmelCase = backbone_class(UpperCAmelCase ) backbone.to(UpperCAmelCase ) backbone.eval() _UpperCAmelCase = backbone(**UpperCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _UpperCAmelCase = backbone(**UpperCAmelCase , output_hidden_states=UpperCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _UpperCAmelCase = backbone(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertIsNotNone(outputs.attentions )
39
0
from ..utils import DummyObject, requires_backends class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Any = ['''sentencepiece'''] def __init__( self : Optional[Any] , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : List[Any] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : str ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : List[str] = ['''sentencepiece'''] def __init__( self : List[str] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : Optional[Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : str = ['''sentencepiece'''] def __init__( self : int , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : int ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : str = ['''sentencepiece'''] def __init__( self : Any , *lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : Any ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Any = ['''sentencepiece'''] def __init__( self : List[Any] , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Optional[int] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : List[str] , *lowerCAmelCase_ : Tuple , **lowerCAmelCase_ : str ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : Any , *lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : int ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Any = ['''sentencepiece'''] def __init__( self : Optional[int] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : List[str] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : Tuple , *lowerCAmelCase_ : Optional[Any] , **lowerCAmelCase_ : Tuple ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Dict = ['''sentencepiece'''] def __init__( self : int , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : List[Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : Optional[int] , *lowerCAmelCase_ : Tuple , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : Optional[Any] , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : List[str] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : Any , *lowerCAmelCase_ : Optional[Any] , **lowerCAmelCase_ : int ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : str , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : List[Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : List[Any] , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : Optional[int] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Any = ['''sentencepiece'''] def __init__( self : Optional[int] , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : int , *lowerCAmelCase_ : int , **lowerCAmelCase_ : Any ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Dict = ['''sentencepiece'''] def __init__( self : Optional[Any] , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : Union[str, Any] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : Tuple ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Union[str, Any] = ['''sentencepiece'''] def __init__( self : Dict , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : str ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : int = ['''sentencepiece'''] def __init__( self : int , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Dict ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : str = ['''sentencepiece'''] def __init__( self : int , *lowerCAmelCase_ : Tuple , **lowerCAmelCase_ : Optional[int] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[Any] = ['''sentencepiece'''] def __init__( self : Any , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : str ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : Dict , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[Any] = ['''sentencepiece'''] def __init__( self : Dict , *lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : List[Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Tuple = ['''sentencepiece'''] def __init__( self : Optional[int] , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : List[str] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Tuple = ['''sentencepiece'''] def __init__( self : Optional[int] , *lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = ['''sentencepiece'''] def __init__( self : Union[str, Any] , *lowerCAmelCase_ : List[str] , **lowerCAmelCase_ : List[str] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Dict = ['''sentencepiece'''] def __init__( self : Union[str, Any] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : Optional[int] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] ) class UpperCAmelCase ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase : Union[str, Any] = ['''sentencepiece'''] def __init__( self : Any , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Optional[Any] ): """simple docstring""" requires_backends(self , ['''sentencepiece'''] )
121
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = TransfoXLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" super().setUp() _UpperCAmelCase = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = '<unk> UNwanted , running' _UpperCAmelCase = '<unk> unwanted, running' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(UpperCAmelCase , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [0, 4, 8, 7] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) _UpperCAmelCase = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' _UpperCAmelCase = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(UpperCAmelCase ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(UpperCAmelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
39
0
"""simple docstring""" import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'''{torch_layer} layer.weight does not match''' _UpperCamelCase = nn.Parameter(__lowerCAmelCase ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'''{torch_layer} layer.bias does not match''' _UpperCamelCase = nn.Parameter(__lowerCAmelCase ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = np.asarray(weights[0] ) _UpperCamelCase = np.asarray(weights[1] ) _UpperCamelCase = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key, torch.tensor(__lowerCAmelCase ).transpose(1, 2 ).contiguous().view(-1, __lowerCAmelCase ), ) set_param( torch_layer.self_attention.value, torch.tensor(__lowerCAmelCase ).transpose(1, 2 ).contiguous().view(-1, __lowerCAmelCase ), ) set_param( torch_layer.output.dense, torch.tensor(__lowerCAmelCase ).view(-1, __lowerCAmelCase ).contiguous().transpose(0, 1 ), ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = np.asarray(weights[0] ) _UpperCamelCase = np.asarray(weights[1] ) _UpperCamelCase = np.asarray(weights[2] ) _UpperCamelCase = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query, torch.tensor(__lowerCAmelCase ).transpose(1, 2 ).contiguous().view(-1, __lowerCAmelCase ), ) set_param( torch_layer.self_attention.key, torch.tensor(__lowerCAmelCase ).transpose(1, 2 ).contiguous().view(-1, __lowerCAmelCase ), ) set_param( torch_layer.self_attention.value, torch.tensor(__lowerCAmelCase ).transpose(1, 2 ).contiguous().view(-1, __lowerCAmelCase ), ) set_param( torch_layer.output.dense, torch.tensor(__lowerCAmelCase ).view(-1, __lowerCAmelCase ).contiguous().transpose(0, 1 ), ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = weights[0][0][0] _UpperCamelCase = np.asarray(layer_norm_a[0] ) _UpperCamelCase = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm, torch.tensor(__lowerCAmelCase ), torch.tensor(__lowerCAmelCase ), ) # lsh weights + output _UpperCamelCase = weights[0][1] if len(__lowerCAmelCase ) < 4: set_layer_weights_in_torch_lsh(__lowerCAmelCase, torch_block.attention, __lowerCAmelCase ) else: set_layer_weights_in_torch_local(__lowerCAmelCase, torch_block.attention, __lowerCAmelCase ) # intermediate weighs _UpperCamelCase = weights[2][0][1][2] # Chunked Feed Forward if len(__lowerCAmelCase ) == 4: _UpperCamelCase = intermediate_weights[2] # layernorm 2 _UpperCamelCase = np.asarray(intermediate_weights[0][0] ) _UpperCamelCase = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm, torch.tensor(__lowerCAmelCase ), torch.tensor(__lowerCAmelCase ), ) # intermediate dense _UpperCamelCase = np.asarray(intermediate_weights[1][0] ) _UpperCamelCase = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense, torch.tensor(__lowerCAmelCase ).transpose(0, 1 ).contiguous(), torch.tensor(__lowerCAmelCase ), ) # intermediate out _UpperCamelCase = np.asarray(intermediate_weights[4][0] ) _UpperCamelCase = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense, torch.tensor(__lowerCAmelCase ).transpose(0, 1 ).contiguous(), torch.tensor(__lowerCAmelCase ), ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = torch_model.reformer # word embeds _UpperCamelCase = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings, torch.tensor(__lowerCAmelCase ), ) if isinstance(weights[3], __lowerCAmelCase ): _UpperCamelCase = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _UpperCamelCase = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'''{position_embeddings[emb_idx]} emb does not match''' _UpperCamelCase = nn.Parameter(torch.tensor(__lowerCAmelCase ) ) _UpperCamelCase = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __lowerCAmelCase ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _UpperCamelCase = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase ) # output layer norm _UpperCamelCase = np.asarray(weights[7][0] ) _UpperCamelCase = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm, torch.tensor(__lowerCAmelCase ), torch.tensor(__lowerCAmelCase ), ) # output embeddings _UpperCamelCase = np.asarray(weights[9][0] ) _UpperCamelCase = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder, torch.tensor(__lowerCAmelCase ).transpose(0, 1 ).contiguous(), torch.tensor(__lowerCAmelCase ), ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = ReformerConfig.from_json_file(__lowerCAmelCase ) print(F'''Building PyTorch model from configuration: {config}''' ) _UpperCamelCase = ReformerModelWithLMHead(__lowerCAmelCase ) with open(__lowerCAmelCase, '''rb''' ) as f: _UpperCamelCase = pickle.load(__lowerCAmelCase )['''weights'''] set_model_weights_in_torch(__lowerCAmelCase, __lowerCAmelCase, config.hidden_size ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict(), __lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _a = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
194
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = {'''configuration_opt''': ['''OPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OPTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''OPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OPTForCausalLM''', '''OPTModel''', '''OPTPreTrainedModel''', '''OPTForSequenceClassification''', '''OPTForQuestionAnswering''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['''TFOPTForCausalLM''', '''TFOPTModel''', '''TFOPTPreTrainedModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''FlaxOPTForCausalLM''', '''FlaxOPTModel''', '''FlaxOPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
39
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ = { '''configuration_roberta''': ['''ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RobertaConfig''', '''RobertaOnnxConfig'''], '''tokenization_roberta''': ['''RobertaTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''RobertaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RobertaForCausalLM''', '''RobertaForMaskedLM''', '''RobertaForMultipleChoice''', '''RobertaForQuestionAnswering''', '''RobertaForSequenceClassification''', '''RobertaForTokenClassification''', '''RobertaModel''', '''RobertaPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRobertaForCausalLM''', '''TFRobertaForMaskedLM''', '''TFRobertaForMultipleChoice''', '''TFRobertaForQuestionAnswering''', '''TFRobertaForSequenceClassification''', '''TFRobertaForTokenClassification''', '''TFRobertaMainLayer''', '''TFRobertaModel''', '''TFRobertaPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''FlaxRobertaForCausalLM''', '''FlaxRobertaForMaskedLM''', '''FlaxRobertaForMultipleChoice''', '''FlaxRobertaForQuestionAnswering''', '''FlaxRobertaForSequenceClassification''', '''FlaxRobertaForTokenClassification''', '''FlaxRobertaModel''', '''FlaxRobertaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
279
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" _UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ), F"""{len(__lowerCAmelCase )} != {len(__lowerCAmelCase )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) _a = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _a = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def __A ( __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" try: _UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(__lowerCAmelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __A ( __lowerCAmelCase , __lowerCAmelCase = "student" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , )-> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" _UpperCAmelCase = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowerCAmelCase , __lowerCAmelCase ): AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) # purely for convenience _UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(__lowerCAmelCase ).eval() else: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), F"""teacher must be a model or string got type {type(__lowerCAmelCase )}""" _UpperCAmelCase = teacher.config.to_diff_dict() try: _UpperCAmelCase , _UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowerCAmelCase ) # Copy weights _UpperCAmelCase = teacher.config_class(**__lowerCAmelCase ) _UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. _UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=__lowerCAmelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save _UpperCAmelCase , _UpperCAmelCase = list(range(__lowerCAmelCase ) ), list(range(__lowerCAmelCase ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(__lowerCAmelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) if d_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) try: if hasattr( __lowerCAmelCase , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowerCAmelCase ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowerCAmelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowerCAmelCase ) copy_layers(teacher.decoder.block , student.decoder.block , __lowerCAmelCase ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) _UpperCAmelCase = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowerCAmelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
39
0
import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEncoder, BertModel, BertPreTrainedModel, ) lowercase__ =logging.getLogger(__name__) class UpperCamelCase__ ( snake_case__ ): def lowerCAmelCase (self : Optional[int] , snake_case_ : List[Any] , snake_case_ : Union[str, Any] , snake_case_ : str=None , snake_case_ : Dict=None ): __a : Optional[int] = self.layer[current_layer](snake_case_ , snake_case_ , head_mask[current_layer] ) __a : Any = layer_outputs[0] return hidden_states @add_start_docstrings( "The bare Bert Model transformer with PABEE outputting raw hidden-states without any specific head on top." ,snake_case__ ,) class UpperCamelCase__ ( snake_case__ ): def __init__(self : Union[str, Any] , snake_case_ : Optional[Any] ): super().__init__(snake_case_ ) __a : Any = BertEncoderWithPabee(snake_case_ ) self.init_weights() __a : Tuple = 0 __a : List[Any] = 0 __a : str = 0 __a : Union[str, Any] = 0 def lowerCAmelCase (self : Tuple , snake_case_ : Any ): __a : int = threshold def lowerCAmelCase (self : Union[str, Any] , snake_case_ : Dict ): __a : Any = patience def lowerCAmelCase (self : Union[str, Any] ): __a : Tuple = 0 __a : Union[str, Any] = 0 def lowerCAmelCase (self : int ): __a : List[str] = self.inference_layers_num / self.inference_instances_num __a : Tuple = ( f"*** Patience = {self.patience} Avg. Inference Layers = {avg_inf_layers:.2f} Speed Up =" f" {1 - avg_inf_layers / self.config.num_hidden_layers:.2f} ***" ) print(snake_case_ ) @add_start_docstrings_to_model_forward(snake_case_ ) def lowerCAmelCase (self : Optional[int] , snake_case_ : Optional[Any]=None , snake_case_ : List[str]=None , snake_case_ : Union[str, Any]=None , snake_case_ : List[Any]=None , snake_case_ : Any=None , snake_case_ : Tuple=None , snake_case_ : Dict=None , snake_case_ : Tuple=None , snake_case_ : List[str]=None , snake_case_ : List[str]=None , snake_case_ : List[str]=False , ): if input_ids is not None and inputs_embeds is not None: raise ValueError('''You cannot specify both input_ids and inputs_embeds at the same time''' ) elif input_ids is not None: __a : List[str] = input_ids.size() elif inputs_embeds is not None: __a : List[Any] = inputs_embeds.size()[:-1] else: raise ValueError('''You have to specify either input_ids or inputs_embeds''' ) __a : Optional[Any] = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: __a : Tuple = torch.ones(snake_case_ , device=snake_case_ ) if token_type_ids is None: __a : Optional[int] = torch.zeros(snake_case_ , dtype=torch.long , device=snake_case_ ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. __a : str = self.get_extended_attention_mask(snake_case_ , snake_case_ , snake_case_ ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if self.config.is_decoder and encoder_hidden_states is not None: __a , __a , __a : str = encoder_hidden_states.size() __a : Optional[Any] = (encoder_batch_size, encoder_sequence_length) if encoder_attention_mask is None: __a : List[Any] = torch.ones(snake_case_ , device=snake_case_ ) __a : List[Any] = self.invert_attention_mask(snake_case_ ) else: __a : int = None # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] __a : Tuple = self.get_head_mask(snake_case_ , self.config.num_hidden_layers ) __a : Union[str, Any] = self.embeddings( input_ids=snake_case_ , position_ids=snake_case_ , token_type_ids=snake_case_ , inputs_embeds=snake_case_ ) __a : Dict = embedding_output if self.training: __a : List[str] = [] for i in range(self.config.num_hidden_layers ): __a : int = self.encoder.adaptive_forward( snake_case_ , current_layer=snake_case_ , attention_mask=snake_case_ , head_mask=snake_case_ ) __a : List[str] = self.pooler(snake_case_ ) __a : Union[str, Any] = output_layers[i](output_dropout(snake_case_ ) ) res.append(snake_case_ ) elif self.patience == 0: # Use all layers for inference __a : Union[str, Any] = self.encoder( snake_case_ , attention_mask=snake_case_ , head_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , ) __a : Optional[Any] = self.pooler(encoder_outputs[0] ) __a : str = [output_layers[self.config.num_hidden_layers - 1](snake_case_ )] else: __a : Dict = 0 __a : Tuple = None __a : Any = 0 for i in range(self.config.num_hidden_layers ): calculated_layer_num += 1 __a : Union[str, Any] = self.encoder.adaptive_forward( snake_case_ , current_layer=snake_case_ , attention_mask=snake_case_ , head_mask=snake_case_ ) __a : Union[str, Any] = self.pooler(snake_case_ ) __a : Tuple = output_layers[i](snake_case_ ) if regression: __a : int = logits.detach() if patient_result is not None: __a : List[str] = patient_result.detach() if (patient_result is not None) and torch.abs(patient_result - labels ) < self.regression_threshold: patient_counter += 1 else: __a : List[Any] = 0 else: __a : List[str] = logits.detach().argmax(dim=1 ) if patient_result is not None: __a : Optional[int] = patient_result.detach().argmax(dim=1 ) if (patient_result is not None) and torch.all(labels.eq(snake_case_ ) ): patient_counter += 1 else: __a : str = 0 __a : List[Any] = logits if patient_counter == self.patience: break __a : List[str] = [patient_result] self.inference_layers_num += calculated_layer_num self.inference_instances_num += 1 return res @add_start_docstrings( "Bert Model transformer with PABEE and a sequence classification/regression head on top (a linear layer on top of\n the pooled output) e.g. for GLUE tasks. " ,snake_case__ ,) class UpperCamelCase__ ( snake_case__ ): def __init__(self : Union[str, Any] , snake_case_ : Optional[Any] ): super().__init__(snake_case_ ) __a : Any = config.num_labels __a : List[Any] = BertModelWithPabee(snake_case_ ) __a : Dict = nn.Dropout(config.hidden_dropout_prob ) __a : str = nn.ModuleList( [nn.Linear(config.hidden_size , self.config.num_labels ) for _ in range(config.num_hidden_layers )] ) self.init_weights() @add_start_docstrings_to_model_forward(snake_case_ ) def lowerCAmelCase (self : Tuple , snake_case_ : Any=None , snake_case_ : Optional[Any]=None , snake_case_ : int=None , snake_case_ : Optional[int]=None , snake_case_ : str=None , snake_case_ : List[Any]=None , snake_case_ : Any=None , ): __a : Optional[Any] = self.bert( input_ids=snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , position_ids=snake_case_ , head_mask=snake_case_ , inputs_embeds=snake_case_ , output_dropout=self.dropout , output_layers=self.classifiers , regression=self.num_labels == 1 , ) __a : Any = (logits[-1],) if labels is not None: __a : Optional[Any] = None __a : Tuple = 0 for ix, logits_item in enumerate(snake_case_ ): if self.num_labels == 1: # We are doing regression __a : Dict = MSELoss() __a : Union[str, Any] = loss_fct(logits_item.view(-1 ) , labels.view(-1 ) ) else: __a : Any = CrossEntropyLoss() __a : List[Any] = loss_fct(logits_item.view(-1 , self.num_labels ) , labels.view(-1 ) ) if total_loss is None: __a : Optional[Any] = loss else: total_loss += loss * (ix + 1) total_weights += ix + 1 __a : str = (total_loss / total_weights,) + outputs return outputs
216
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False )-> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase = '' else: _UpperCAmelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase = in_proj_bias[: config.hidden_size] _UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase = in_proj_bias[-config.hidden_size :] def __A ( __lowerCAmelCase )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" _UpperCAmelCase = dct.pop(__lowerCAmelCase ) _UpperCAmelCase = val def __A ( )-> str: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True )-> List[str]: """simple docstring""" _UpperCAmelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCAmelCase = 8 # set labels if required if not base_model: _UpperCAmelCase = 1_000 _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCAmelCase = 384 _UpperCAmelCase = 1_536 _UpperCAmelCase = 12 _UpperCAmelCase = 6 # load original model from torch hub _UpperCAmelCase = torch.hub.load('facebookresearch/dino:main' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCAmelCase = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _UpperCAmelCase = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _UpperCAmelCase = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCAmelCase = ViTImageProcessor() _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCAmelCase = encoding['pixel_values'] _UpperCAmelCase = model(__lowerCAmelCase ) if base_model: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) _a = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
39
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def UpperCamelCase_( _snake_case : str ): """simple docstring""" __a =[tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __magic_name__ ( snake_case__ , snake_case__ , snake_case__ , unittest.TestCase ): SCREAMING_SNAKE_CASE = StableDiffusionLatentUpscalePipeline SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'height', 'width', 'cross_attention_kwargs', 'negative_prompt_embeds', 'prompt_embeds', } SCREAMING_SNAKE_CASE = PipelineTesterMixin.required_optional_params - {'num_images_per_prompt'} SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS SCREAMING_SNAKE_CASE = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess SCREAMING_SNAKE_CASE = frozenset([] ) SCREAMING_SNAKE_CASE = True @property def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =1 __a =4 __a =(16, 16) __a =floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__snake_case ) return image def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) __a =UNetaDConditionModel( act_fn='gelu' , attention_head_dim=8 , norm_num_groups=__snake_case , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=160 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( 'KDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', ) , in_channels=8 , mid_block_type=__snake_case , only_cross_attention=__snake_case , out_channels=5 , resnet_time_scale_shift='scale_shift' , time_embedding_type='fourier' , timestep_post_act='gelu' , up_block_types=('KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KUpBlock2D') , ) __a =AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', ] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) __a =EulerDiscreteScheduler(prediction_type='sample' ) __a =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='quick_gelu' , projection_dim=512 , ) __a =CLIPTextModel(__snake_case ) __a =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __a ={ 'unet': model.eval(), 'vae': vae.eval(), 'scheduler': scheduler, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def __magic_name__ ( self , __snake_case , __snake_case=0 ) -> Union[str, Any]: '''simple docstring''' if str(__snake_case ).startswith('mps' ): __a =torch.manual_seed(__snake_case ) else: __a =torch.Generator(device=__snake_case ).manual_seed(__snake_case ) __a ={ 'prompt': 'A painting of a squirrel eating a burger', 'image': self.dummy_image.cpu(), 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' __a ='cpu' __a =self.get_dummy_components() __a =self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) __a =self.get_dummy_inputs(__snake_case ) __a =pipe(**__snake_case ).images __a =image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 256, 256, 3) ) __a =np.array( [0.4722_2412, 0.4192_1633, 0.4471_7434, 0.4687_4192, 0.4258_8258, 0.4615_0726, 0.467_7534, 0.4558_3832, 0.4857_9055] ) __a =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__snake_case , 1e-3 ) def __magic_name__ ( self ) -> Dict: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def __magic_name__ ( self ) -> int: '''simple docstring''' super().test_save_load_local(expected_max_difference=3e-3 ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3e-3 ) def __magic_name__ ( self ) -> Tuple: '''simple docstring''' __a =[ 'DDIMScheduler', 'DDPMScheduler', 'PNDMScheduler', 'HeunDiscreteScheduler', 'EulerAncestralDiscreteScheduler', 'KDPM2DiscreteScheduler', 'KDPM2AncestralDiscreteScheduler', 'DPMSolverSDEScheduler', ] __a =self.get_dummy_components() __a =self.pipeline_class(**__snake_case ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) __a =self.get_dummy_inputs(__snake_case ) __a =2 __a =[] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue __a =getattr(__snake_case , scheduler_enum.name ) __a =scheduler_cls.from_config(pipe.scheduler.config ) __a =pipe(**__snake_case )[0] outputs.append(__snake_case ) assert check_same_shape(__snake_case ) @require_torch_gpu @slow class __magic_name__ ( unittest.TestCase ): def __magic_name__ ( self ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self ) -> int: '''simple docstring''' __a =torch.manual_seed(33 ) __a =StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' , torch_dtype=torch.floataa ) pipe.to('cuda' ) __a =StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) __a ='a photo of an astronaut high resolution, unreal engine, ultra realistic' __a =pipe(__snake_case , generator=__snake_case , output_type='latent' ).images __a =upscaler( prompt=__snake_case , image=__snake_case , num_inference_steps=20 , guidance_scale=0 , generator=__snake_case , output_type='np' , ).images[0] __a =load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy' ) assert np.abs((expected_image - image).mean() ) < 5e-2 def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a =torch.manual_seed(33 ) __a =StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) __a ='the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas' __a =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png' ) __a =upscaler( prompt=__snake_case , image=__snake_case , num_inference_steps=20 , guidance_scale=0 , generator=__snake_case , output_type='np' , ).images[0] __a =load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy' ) assert np.abs((expected_image - image).max() ) < 5e-2
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _snake_case = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class _snake_case ( unittest.TestCase ): lowerCamelCase__: int = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowerCamelCase__: Tuple = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: lowerCamelCase__: Union[str, Any] = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: lowerCamelCase__: List[str] = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _lowerCamelCase ( self: Any , __lowerCamelCase: Dict , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Optional[Any] ) -> List[str]: __UpperCAmelCase : Dict = ZeroShotClassificationPipeline( model=__lowerCamelCase , tokenizer=__lowerCamelCase , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _lowerCamelCase ( self: Tuple , __lowerCamelCase: Tuple , __lowerCamelCase: str ) -> Optional[int]: __UpperCAmelCase : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(__lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase )]} ) # No kwarg __UpperCAmelCase : Optional[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(__lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase )]} ) __UpperCAmelCase : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(__lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase )]} ) __UpperCAmelCase : Any = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( __lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) __UpperCAmelCase : int = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( __lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) __UpperCAmelCase : Tuple = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(__lowerCamelCase , {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase )]} ) # https://github.com/huggingface/transformers/issues/13846 __UpperCAmelCase : List[str] = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( __lowerCamelCase , [ {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )]} for i in range(1 ) ] , ) __UpperCAmelCase : List[Any] = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( __lowerCamelCase , [ {"sequence": ANY(__lowerCamelCase ), "labels": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )], "scores": [ANY(__lowerCamelCase ), ANY(__lowerCamelCase )]} for i in range(2 ) ] , ) with self.assertRaises(__lowerCamelCase ): classifier("" , candidate_labels="politics" ) with self.assertRaises(__lowerCamelCase ): classifier(__lowerCamelCase , candidate_labels="politics" ) with self.assertRaises(__lowerCamelCase ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(__lowerCamelCase ): classifier("Who are you voting for in 2020?" , candidate_labels=__lowerCamelCase ) with self.assertRaises(__lowerCamelCase ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(__lowerCamelCase ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=__lowerCamelCase , ) self.run_entailment_id(__lowerCamelCase ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: Optional[int] ) -> Optional[int]: __UpperCAmelCase : Dict = zero_shot_classifier.model.config __UpperCAmelCase : List[Any] = config.labelaid __UpperCAmelCase : str = zero_shot_classifier.entailment_id __UpperCAmelCase : Dict = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) __UpperCAmelCase : Any = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) __UpperCAmelCase : Dict = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) __UpperCAmelCase : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) __UpperCAmelCase : int = original_labelaid self.assertEqual(__lowerCamelCase , zero_shot_classifier.entailment_id ) @require_torch def _lowerCamelCase ( self: Optional[Any] ) -> List[Any]: __UpperCAmelCase : Optional[int] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 1_00 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _lowerCamelCase ( self: str ) -> Any: __UpperCAmelCase : int = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) __UpperCAmelCase : Union[str, Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_33, 0.3_33, 0.3_33], } , ) @require_tf def _lowerCamelCase ( self: Tuple ) -> Tuple: __UpperCAmelCase : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) __UpperCAmelCase : Tuple = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_33, 0.3_33, 0.3_33], } , ) @slow @require_torch def _lowerCamelCase ( self: Any ) -> List[str]: __UpperCAmelCase : List[str] = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) __UpperCAmelCase : Dict = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_76, 0.0_15, 0.0_09], } , ) __UpperCAmelCase : Optional[Any] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=__lowerCamelCase , ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , ) @slow @require_tf def _lowerCamelCase ( self: List[Any] ) -> Any: __UpperCAmelCase : Tuple = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) __UpperCAmelCase : List[str] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_76, 0.0_15, 0.0_09], } , ) __UpperCAmelCase : Optional[Any] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=__lowerCamelCase , ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , )
157
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=3 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=10 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[1, 1, 2, 1] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase="relu" , UpperCAmelCase=3 , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = embeddings_size _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = scope _UpperCAmelCase = len(UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = TFResNetModel(config=UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFResNetForImageClassification(UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase = layer_type _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) @slow def UpperCamelCase ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TFResNetModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __A ( )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" @cached_property def UpperCamelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=UpperCAmelCase , return_tensors='tf' ) # forward pass _UpperCAmelCase = model(**UpperCAmelCase ) # verify the logits _UpperCAmelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _UpperCAmelCase = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCAmelCase , atol=1e-4 ) )
39
0
'''simple docstring''' import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) class lowerCAmelCase_ ( enum.Enum ): __lowerCamelCase : List[str] = 0 __lowerCamelCase : int = 1 @add_end_docstrings(snake_case__ ) class lowerCAmelCase_ ( snake_case__ ): __lowerCamelCase : str = "generated" def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ) -> int: super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def _snake_case ( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase , ) -> Union[str, Any]: _lowerCAmelCase = {} if truncation is not None: _lowerCAmelCase = truncation _lowerCAmelCase = generate_kwargs _lowerCAmelCase = {} if return_tensors is not None and return_type is None: _lowerCAmelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: _lowerCAmelCase = return_type if clean_up_tokenization_spaces is not None: _lowerCAmelCase = clean_up_tokenization_spaces if stop_sequence is not None: _lowerCAmelCase = self.tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) if len(_lowerCAmelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) _lowerCAmelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: return True def _snake_case ( self , *_lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: _lowerCAmelCase = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , _lowerCAmelCase ): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input" ) _lowerCAmelCase = ([prefix + arg for arg in args[0]],) _lowerCAmelCase = True elif isinstance(args[0] , _lowerCAmelCase ): _lowerCAmelCase = (prefix + args[0],) _lowerCAmelCase = False else: raise ValueError( f''' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`''' ) _lowerCAmelCase = self.tokenizer(*_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self , *_lowerCAmelCase , **_lowerCAmelCase ) -> Optional[Any]: _lowerCAmelCase = super().__call__(*_lowerCAmelCase , **_lowerCAmelCase ) if ( isinstance(args[0] , _lowerCAmelCase ) and all(isinstance(_lowerCAmelCase , _lowerCAmelCase ) for el in args[0] ) and all(len(_lowerCAmelCase ) == 1 for res in result ) ): return [res[0] for res in result] return result def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase=TruncationStrategy.DO_NOT_TRUNCATE , **_lowerCAmelCase ) -> str: _lowerCAmelCase = self._parse_and_tokenize(_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase ) return inputs def _snake_case ( self , _lowerCAmelCase , **_lowerCAmelCase ) -> List[str]: if self.framework == "pt": _lowerCAmelCase , _lowerCAmelCase = model_inputs["input_ids"].shape elif self.framework == "tf": _lowerCAmelCase , _lowerCAmelCase = tf.shape(model_inputs["input_ids"] ).numpy() _lowerCAmelCase = generate_kwargs.get("min_length" , self.model.config.min_length ) _lowerCAmelCase = generate_kwargs.get("max_length" , self.model.config.max_length ) self.check_inputs(_lowerCAmelCase , generate_kwargs["min_length"] , generate_kwargs["max_length"] ) _lowerCAmelCase = self.model.generate(**_lowerCAmelCase , **_lowerCAmelCase ) _lowerCAmelCase = output_ids.shape[0] if self.framework == "pt": _lowerCAmelCase = output_ids.reshape(_lowerCAmelCase , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": _lowerCAmelCase = tf.reshape(_lowerCAmelCase , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase=ReturnType.TEXT , _lowerCAmelCase=False ) -> int: _lowerCAmelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: _lowerCAmelCase = {f'''{self.return_name}_token_ids''': output_ids} elif return_type == ReturnType.TEXT: _lowerCAmelCase = { f'''{self.return_name}_text''': self.tokenizer.decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase , ) } records.append(_lowerCAmelCase ) return records @add_end_docstrings(snake_case__ ) class lowerCAmelCase_ ( snake_case__ ): __lowerCamelCase : int = "summary" def __call__( self , *_lowerCAmelCase , **_lowerCAmelCase ) -> List[str]: return super().__call__(*_lowerCAmelCase , **_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[str]: if max_length < min_length: logger.warning(f'''Your min_length={min_length} must be inferior than your max_length={max_length}.''' ) if input_length < max_length: logger.warning( f'''Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ''' "a summarization task, where outputs shorter than the input are typically wanted, you might " f'''consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})''' ) @add_end_docstrings(snake_case__ ) class lowerCAmelCase_ ( snake_case__ ): __lowerCamelCase : Union[str, Any] = "translation" def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> str: if input_length > 0.9 * max_length: logger.warning( f'''Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ''' "increasing your max_length manually, e.g. translator(\'...\', max_length=400)" ) return True def _snake_case ( self , *_lowerCAmelCase , _lowerCAmelCase=TruncationStrategy.DO_NOT_TRUNCATE , _lowerCAmelCase=None , _lowerCAmelCase=None ) -> str: if getattr(self.tokenizer , "_build_translation_inputs" , _lowerCAmelCase ): return self.tokenizer._build_translation_inputs( *_lowerCAmelCase , return_tensors=self.framework , truncation=_lowerCAmelCase , src_lang=_lowerCAmelCase , tgt_lang=_lowerCAmelCase ) else: return super()._parse_and_tokenize(*_lowerCAmelCase , truncation=_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ) -> List[Any]: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = super()._sanitize_parameters(**_lowerCAmelCase ) if src_lang is not None: _lowerCAmelCase = src_lang if tgt_lang is not None: _lowerCAmelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. _lowerCAmelCase = kwargs.get("task" , self.task ) _lowerCAmelCase = task.split("_" ) if task and len(_lowerCAmelCase ) == 4: # translation, XX, to YY _lowerCAmelCase = items[1] _lowerCAmelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self , *_lowerCAmelCase , **_lowerCAmelCase ) -> Optional[Any]: return super().__call__(*_lowerCAmelCase , **_lowerCAmelCase )
158
def __A ( __lowerCAmelCase )-> list: """simple docstring""" if len(__lowerCAmelCase ) < 2: return collection def circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , mid + 1 , __lowerCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) - 1 ) return collection if __name__ == "__main__": _a = input('''Enter numbers separated by a comma:\n''').strip() _a = [int(item) for item in user_input.split(''',''')] print(circle_sort(unsorted))
39
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class A__ ( snake_case__ ): lowercase = ( 'This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.' 'It takes two arguments named `image` which should be the original image, and `label` which should be a text ' 'describing the elements what should be identified in the segmentation mask. The tool returns the mask.' ) lowercase = 'CIDAS/clipseg-rd64-refined' lowercase = 'image_segmenter' lowercase = CLIPSegForImageSegmentation lowercase = ['image', 'text'] lowercase = ['image'] def __init__( self : Optional[int] , *a : Any , **a : Optional[int] ): '''simple docstring''' requires_backends(self , ['vision'] ) super().__init__(*a , **a ) def _lowerCamelCase ( self : str , a : Any , a : List[str] ): '''simple docstring''' return self.pre_processor(text=[label] , images=[image] , padding=a , return_tensors='pt' ) def _lowerCamelCase ( self : str , a : List[Any] ): '''simple docstring''' with torch.no_grad(): lowerCAmelCase__ : Tuple = self.model(**a ).logits return logits def _lowerCamelCase ( self : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : Any = outputs.cpu().detach().numpy() lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : Tuple = 1 return Image.fromarray((array * 255).astype(np.uinta ) )
212
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "Pix2StructImageProcessor" UpperCamelCase__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = False super().__init__(UpperCAmelCase , UpperCAmelCase ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 2048 , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , **UpperCAmelCase ) else: # add pixel_values and bbox _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , header_text=UpperCAmelCase , **UpperCAmelCase ) if text is not None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) if "attention_mask" in text_encoding: _UpperCAmelCase = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: _UpperCAmelCase = text_encoding.pop('input_ids' ) else: _UpperCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase ) return encoding_image_processor def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
from __future__ import annotations import collections import pprint from pathlib import Path def UpperCAmelCase ( a_ ) -> str: """simple docstring""" return "".join(sorted(__lowerCAmelCase ) ) def UpperCAmelCase ( a_ ) -> list[str]: """simple docstring""" return word_by_signature[signature(__lowerCAmelCase )] SCREAMING_SNAKE_CASE :List[Any] = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') SCREAMING_SNAKE_CASE :Any = sorted({word.strip().lower() for word in data.splitlines()}) SCREAMING_SNAKE_CASE :Optional[Any] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": SCREAMING_SNAKE_CASE :Optional[Any] = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
15
class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase = "" , UpperCAmelCase = False ): """simple docstring""" _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 0 for q, w in zip(self.prefix , UpperCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=UpperCAmelCase , is_leaf=UpperCAmelCase ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def UpperCamelCase ( self , UpperCAmelCase = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = RadixNode() root.insert_many(__lowerCAmelCase ) assert all(root.find(__lowerCAmelCase ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = 'banana bananas bandanas bandana band apple all beast'.split() root.insert_many(__lowerCAmelCase ) print('Words:' , __lowerCAmelCase ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
39
0
def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Tuple ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = [0] * len(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__lowerCAmelCase ) ): if indegree[i] == 0: queue.append(__lowerCAmelCase ) while queue: SCREAMING_SNAKE_CASE__ = queue.pop(0 ) cnt += 1 topo.append(__lowerCAmelCase ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__lowerCAmelCase ) if cnt != len(__lowerCAmelCase ): print("""Cycle exists""" ) else: print(__lowerCAmelCase ) # Adjacency List of Graph __lowerCamelCase : Tuple = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
219
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class __lowerCamelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments UpperCAmelCase="<s>" , UpperCAmelCase="<pad>" , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = bos, unk, pad, eos _UpperCAmelCase = [] _UpperCAmelCase = [] _UpperCAmelCase = {} _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , UpperCAmelCase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , UpperCAmelCase ): """simple docstring""" return sym in self.indices @classmethod def UpperCamelCase ( cls , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = cls() d.add_from_file(UpperCAmelCase ) return d def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase=1 , UpperCAmelCase=False ): """simple docstring""" if word in self.indices and not overwrite: _UpperCAmelCase = self.indices[word] _UpperCAmelCase = self.count[idx] + n return idx else: _UpperCAmelCase = len(self.symbols ) _UpperCAmelCase = idx self.symbols.append(UpperCAmelCase ) self.count.append(UpperCAmelCase ) return idx def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return 0 def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(UpperCAmelCase ) ) return _UpperCAmelCase = f.readlines() _UpperCAmelCase = self._load_meta(UpperCAmelCase ) for line in lines[indices_start_line:]: try: _UpperCAmelCase , _UpperCAmelCase = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": _UpperCAmelCase = True _UpperCAmelCase , _UpperCAmelCase = line.rsplit(' ' , 1 ) else: _UpperCAmelCase = False _UpperCAmelCase = int(UpperCAmelCase ) _UpperCAmelCase = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(UpperCAmelCase ) ) self.add_symbol(UpperCAmelCase , n=UpperCAmelCase , overwrite=UpperCAmelCase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" _UpperCAmelCase = dict((re.sub(R'@@$' , '' , __lowerCAmelCase ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , __lowerCAmelCase ), v) for k, v in d.items() ) _UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] _UpperCAmelCase = d[k] # restore return da def __A ( __lowerCAmelCase , __lowerCAmelCase )-> str: """simple docstring""" if not os.path.exists(__lowerCAmelCase ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'checkpoint.pt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location='cpu' ) _UpperCAmelCase = chkpt['cfg']['model'] # dicts _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'dict.txt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) _UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) _UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) _UpperCAmelCase = len(__lowerCAmelCase ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['vocab_file'] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # merges_file (bpecodes) _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'bpecodes' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowerCAmelCase , __lowerCAmelCase ) # model config _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'config.json' ) _UpperCAmelCase = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1E-12, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # tokenizer config _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 1_024, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # model _UpperCAmelCase = chkpt['model'] # remove unneeded keys _UpperCAmelCase = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) else: _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase = BioGptConfig.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase = BioGptForCausalLM(__lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(__lowerCAmelCase ) # save _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) print('Conversion is done!' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
39
0
import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class a__ ( snake_case__ ): """simple docstring""" __lowerCamelCase = 'MCTCTFeatureExtractor' __lowerCamelCase = 'AutoTokenizer' def __init__( self , lowercase , lowercase ) -> str: '''simple docstring''' super().__init__(lowercase , lowercase ) A__ = self.feature_extractor A__ = False def __call__( self , *lowercase , **lowercase ) -> Optional[Any]: '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*lowercase , **lowercase ) if "raw_speech" in kwargs: warnings.warn("Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead." ) A__ = kwargs.pop("raw_speech" ) else: A__ = kwargs.pop("audio" , lowercase ) A__ = kwargs.pop("sampling_rate" , lowercase ) A__ = kwargs.pop("text" , lowercase ) if len(lowercase ) > 0: A__ = args[0] A__ = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if audio is not None: A__ = self.feature_extractor(lowercase , *lowercase , sampling_rate=lowercase , **lowercase ) if text is not None: A__ = self.tokenizer(lowercase , **lowercase ) if text is None: return inputs elif audio is None: return encodings else: A__ = encodings["input_ids"] return inputs def UpperCamelCase ( self , *lowercase , **lowercase ) -> Optional[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*lowercase , **lowercase ) def UpperCamelCase ( self , *lowercase , **lowercase ) -> Union[str, Any]: '''simple docstring''' if self._in_target_context_manager: return self.current_processor.pad(*lowercase , **lowercase ) A__ = kwargs.pop("input_features" , lowercase ) A__ = kwargs.pop("labels" , lowercase ) if len(lowercase ) > 0: A__ = args[0] A__ = args[1:] if input_features is not None: A__ = self.feature_extractor.pad(lowercase , *lowercase , **lowercase ) if labels is not None: A__ = self.tokenizer.pad(lowercase , **lowercase ) if labels is None: return input_features elif input_features is None: return labels else: A__ = labels["input_ids"] return input_features def UpperCamelCase ( self , *lowercase , **lowercase ) -> str: '''simple docstring''' return self.tokenizer.decode(*lowercase , **lowercase ) @contextmanager def UpperCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' warnings.warn( "`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your " "labels by using the argument `text` of the regular `__call__` method (either in the same call as " "your audio inputs, or in a separate call." ) A__ = True A__ = self.tokenizer yield A__ = self.feature_extractor A__ = False
68
from __future__ import annotations import collections import pprint from pathlib import Path def __A ( __lowerCAmelCase )-> str: """simple docstring""" return "".join(sorted(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase )-> list[str]: """simple docstring""" return word_by_signature[signature(__lowerCAmelCase )] _a = Path(__file__).parent.joinpath('''words.txt''').read_text(encoding='''utf-8''') _a = sorted({word.strip().lower() for word in data.splitlines()}) _a = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _a = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('''anagrams.txt''', '''w''') as file: file.write('''all_anagrams = \n ''') file.write(pprint.pformat(all_anagrams))
39
0
# DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class UpperCAmelCase ( snake_case__ ): '''simple docstring''' __UpperCamelCase : Tuple = 42 __UpperCamelCase : Optional[Any] = None def lowerCamelCase__ ( a , a=0.999 , a="cosine" , ) -> Union[str, Any]: if alpha_transform_type == "cosine": def alpha_bar_fn(a ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(a ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) _A: List[str] = [] for i in range(__lowerCAmelCase ): _A: Union[str, Any] = i / num_diffusion_timesteps _A: str = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__lowerCAmelCase ) / alpha_bar_fn(__lowerCAmelCase ) , __lowerCAmelCase ) ) return torch.tensor(__lowerCAmelCase , dtype=torch.floataa ) class UpperCAmelCase ( snake_case__ , snake_case__ ): '''simple docstring''' __UpperCamelCase : Dict = 1 @register_to_config def __init__( self : List[Any] , lowerCAmelCase_ : Optional[int] = 1_0_0_0 , lowerCAmelCase_ : List[Any] = 0.0001 , lowerCAmelCase_ : Union[str, Any] = 0.02 , lowerCAmelCase_ : Optional[int] = "linear" , lowerCAmelCase_ : List[str] = None , lowerCAmelCase_ : Any = True , lowerCAmelCase_ : Union[str, Any] = True , lowerCAmelCase_ : str = 0 , lowerCAmelCase_ : List[str] = "epsilon" , lowerCAmelCase_ : Any = 1.0 , **lowerCAmelCase_ : List[Any] , ): """simple docstring""" if kwargs.get('''set_alpha_to_one''' , lowerCAmelCase_ ) is not None: _A: Tuple = ( '''The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.''' ) deprecate('''set_alpha_to_one''' , '''1.0.0''' , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) _A: Dict = kwargs['''set_alpha_to_one'''] if trained_betas is not None: _A: Dict = torch.tensor(lowerCAmelCase_ , dtype=torch.floataa ) elif beta_schedule == "linear": _A: Any = torch.linspace(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _A: Optional[Any] = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , lowerCAmelCase_ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _A: Any = betas_for_alpha_bar(lowerCAmelCase_ ) else: raise NotImplementedError(F"""{beta_schedule} does is not implemented for {self.__class__}""" ) _A: Any = 1.0 - self.betas _A: Union[str, Any] = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. _A: Dict = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution _A: List[str] = 1.0 # setable values _A: int = None _A: Any = torch.from_numpy(np.arange(0 , lowerCAmelCase_ ).copy().astype(np.intaa ) ) def __magic_name__ ( self : List[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple = None ): """simple docstring""" return sample def __magic_name__ ( self : Any , lowerCAmelCase_ : Any , lowerCAmelCase_ : Union[str, Any] = None ): """simple docstring""" if num_inference_steps > self.config.num_train_timesteps: raise ValueError( F"""`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:""" F""" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle""" F""" maximal {self.config.num_train_timesteps} timesteps.""" ) _A: Any = num_inference_steps _A: Any = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 _A: Union[str, Any] = (np.arange(0 , lowerCAmelCase_ ) * step_ratio).round().copy().astype(np.intaa ) _A: int = torch.from_numpy(lowerCAmelCase_ ).to(lowerCAmelCase_ ) self.timesteps += self.config.steps_offset def __magic_name__ ( self : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[Any] = 0.0 , lowerCAmelCase_ : List[str] = False , lowerCAmelCase_ : Union[str, Any] = None , lowerCAmelCase_ : Dict = True , ): """simple docstring""" _A: List[Any] = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process _A: Union[str, Any] = self.alphas_cumprod[timestep] _A: Tuple = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) _A: List[str] = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": _A: Any = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 _A: Tuple = model_output elif self.config.prediction_type == "sample": _A: Any = model_output _A: Dict = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": _A: int = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output _A: Optional[Any] = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or""" ''' `v_prediction`''' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: _A: Tuple = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _A: int = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _A: Union[str, Any] = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=lowerCAmelCase_ , pred_original_sample=lowerCAmelCase_ ) def __len__( self : Optional[int] ): """simple docstring""" return self.config.num_train_timesteps
121
from __future__ import annotations def __A ( __lowerCAmelCase )-> list[int]: """simple docstring""" _UpperCAmelCase = 2 _UpperCAmelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCAmelCase ) if n > 1: factors.append(__lowerCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
39
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/config.json""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/config.json""", """funnel-transformer/medium-base""": """https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json""", """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/config.json""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json""", """funnel-transformer/xlarge-base""": """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json""", } class _UpperCAmelCase( snake_case__ ): lowercase__ = 'funnel' lowercase__ = { 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', } def __init__( self , __a=3_05_22 , __a=[4, 4, 4] , __a=None , __a=2 , __a=7_68 , __a=12 , __a=64 , __a=30_72 , __a="gelu_new" , __a=0.1 , __a=0.1 , __a=0.0 , __a=0.1 , __a=None , __a=1e-9 , __a="mean" , __a="relative_shift" , __a=True , __a=True , __a=True , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = block_sizes _UpperCamelCase = [1] * len(__a) if block_repeats is None else block_repeats assert len(__a) == len( self.block_repeats), "`block_sizes` and `block_repeats` should have the same length." _UpperCamelCase = num_decoder_layers _UpperCamelCase = d_model _UpperCamelCase = n_head _UpperCamelCase = d_head _UpperCamelCase = d_inner _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = initializer_range _UpperCamelCase = initializer_std _UpperCamelCase = layer_norm_eps assert pooling_type in [ "mean", "max", ], F'''Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.''' _UpperCamelCase = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F'''Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.''' _UpperCamelCase = attention_type _UpperCamelCase = separate_cls _UpperCamelCase = truncate_seq _UpperCamelCase = pool_q_only super().__init__(**__a) @property def UpperCAmelCase ( self) -> Any: '''simple docstring''' return sum(self.block_sizes) @num_hidden_layers.setter def UpperCAmelCase ( self , __a) -> Dict: '''simple docstring''' raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.''') @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return len(self.block_sizes) @num_blocks.setter def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' raise NotImplementedError('''This model does not support the setting of `num_blocks`. Please set `block_sizes`.''')
194
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __A ( )-> tuple[list[int], int]: """simple docstring""" _UpperCAmelCase = [randint(-1_000 , 1_000 ) for i in range(10 )] _UpperCAmelCase = randint(-5_000 , 5_000 ) return (arr, r) _a = make_dataset() def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, ...]: """simple docstring""" for triplet in permutations(__lowerCAmelCase , 3 ): if sum(__lowerCAmelCase ) == target: return tuple(sorted(__lowerCAmelCase ) ) return (0, 0, 0) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, int, int]: """simple docstring""" arr.sort() _UpperCAmelCase = len(__lowerCAmelCase ) for i in range(n - 1 ): _UpperCAmelCase , _UpperCAmelCase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __A ( )-> tuple[float, float]: """simple docstring""" _UpperCAmelCase = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' _UpperCAmelCase = '\ntriplet_sum1(*dataset)\n' _UpperCAmelCase = '\ntriplet_sum2(*dataset)\n' _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) return (min(__lowerCAmelCase ), min(__lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _a = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
39
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ = logging.get_logger(__name__) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase=False ) -> Union[str, Any]: """simple docstring""" snake_case_ : Tuple = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('''head''' ): snake_case_ : Union[str, Any] = '''segformer.encoder.''' + key if key.startswith('''backbone''' ): snake_case_ : Any = key.replace('''backbone''' , '''segformer.encoder''' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 snake_case_ : str = key[key.find('''patch_embed''' ) + len('''patch_embed''' )] snake_case_ : List[str] = key.replace(f'''patch_embed{idx}''' , f'''patch_embeddings.{int(__lowerCAmelCase )-1}''' ) if "norm" in key: snake_case_ : Any = key.replace('''norm''' , '''layer_norm''' ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 snake_case_ : int = key[key.find('''segformer.encoder.layer_norm''' ) + len('''segformer.encoder.layer_norm''' )] snake_case_ : Dict = key.replace(f'''layer_norm{idx}''' , f'''layer_norm.{int(__lowerCAmelCase )-1}''' ) if "layer_norm1" in key: snake_case_ : Dict = key.replace('''layer_norm1''' , '''layer_norm_1''' ) if "layer_norm2" in key: snake_case_ : Optional[int] = key.replace('''layer_norm2''' , '''layer_norm_2''' ) if "block" in key: # replace for example block1 by block.0 snake_case_ : List[Any] = key[key.find('''block''' ) + len('''block''' )] snake_case_ : Union[str, Any] = key.replace(f'''block{idx}''' , f'''block.{int(__lowerCAmelCase )-1}''' ) if "attn.q" in key: snake_case_ : Optional[int] = key.replace('''attn.q''' , '''attention.self.query''' ) if "attn.proj" in key: snake_case_ : Union[str, Any] = key.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in key: snake_case_ : Dict = key.replace('''attn''' , '''attention.self''' ) if "fc1" in key: snake_case_ : Optional[int] = key.replace('''fc1''' , '''dense1''' ) if "fc2" in key: snake_case_ : Tuple = key.replace('''fc2''' , '''dense2''' ) if "linear_pred" in key: snake_case_ : Any = key.replace('''linear_pred''' , '''classifier''' ) if "linear_fuse" in key: snake_case_ : str = key.replace('''linear_fuse.conv''' , '''linear_fuse''' ) snake_case_ : Tuple = key.replace('''linear_fuse.bn''' , '''batch_norm''' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 snake_case_ : Tuple = key[key.find('''linear_c''' ) + len('''linear_c''' )] snake_case_ : List[str] = key.replace(f'''linear_c{idx}''' , f'''linear_c.{int(__lowerCAmelCase )-1}''' ) if key.startswith('''head''' ): snake_case_ : List[str] = key.replace('''head''' , '''classifier''' ) snake_case_ : str = value return new_state_dict def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> List[str]: """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) snake_case_ : Optional[int] = state_dict.pop(f'''segformer.encoder.block.{i}.{j}.attention.self.kv.weight''' ) snake_case_ : int = state_dict.pop(f'''segformer.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict snake_case_ : str = kv_weight[ : config.hidden_sizes[i], : ] snake_case_ : List[str] = kv_bias[: config.hidden_sizes[i]] snake_case_ : List[str] = kv_weight[ config.hidden_sizes[i] :, : ] snake_case_ : int = kv_bias[ config.hidden_sizes[i] : ] def lowerCamelCase_ ( ) -> Union[str, Any]: """simple docstring""" snake_case_ : Dict = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case_ : Dict = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return image @torch.no_grad() def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Tuple: """simple docstring""" snake_case_ : Optional[int] = SegformerConfig() snake_case_ : List[Any] = False # set attributes based on model_name snake_case_ : List[Any] = '''huggingface/label-files''' if "segformer" in model_name: snake_case_ : Tuple = model_name[len('''segformer.''' ) : len('''segformer.''' ) + 2] if "ade" in model_name: snake_case_ : str = 150 snake_case_ : Optional[Any] = '''ade20k-id2label.json''' snake_case_ : int = (1, 150, 128, 128) elif "city" in model_name: snake_case_ : Any = 19 snake_case_ : Tuple = '''cityscapes-id2label.json''' snake_case_ : List[Any] = (1, 19, 128, 128) else: raise ValueError(f'''Model {model_name} not supported''' ) elif "mit" in model_name: snake_case_ : List[Any] = True snake_case_ : List[str] = model_name[4:6] snake_case_ : Dict = 1_000 snake_case_ : Optional[Any] = '''imagenet-1k-id2label.json''' snake_case_ : Optional[Any] = (1, 1_000) else: raise ValueError(f'''Model {model_name} not supported''' ) # set config attributes snake_case_ : Tuple = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case_ : Optional[int] = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} snake_case_ : List[str] = idalabel snake_case_ : List[str] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": snake_case_ : str = [64, 128, 320, 512] snake_case_ : Optional[Any] = 256 elif size == "b2": snake_case_ : List[Any] = [64, 128, 320, 512] snake_case_ : Union[str, Any] = 768 snake_case_ : int = [3, 4, 6, 3] elif size == "b3": snake_case_ : List[str] = [64, 128, 320, 512] snake_case_ : Optional[int] = 768 snake_case_ : int = [3, 4, 18, 3] elif size == "b4": snake_case_ : Dict = [64, 128, 320, 512] snake_case_ : List[Any] = 768 snake_case_ : Tuple = [3, 8, 27, 3] elif size == "b5": snake_case_ : Dict = [64, 128, 320, 512] snake_case_ : Tuple = 768 snake_case_ : List[Any] = [3, 6, 40, 3] else: raise ValueError(f'''Size {size} not supported''' ) # load image processor (only resize + normalize) snake_case_ : Optional[Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__lowerCAmelCase , align=__lowerCAmelCase , do_random_crop=__lowerCAmelCase ) # prepare image snake_case_ : Optional[Any] = prepare_img() snake_case_ : List[Any] = image_processor(images=__lowerCAmelCase , return_tensors='''pt''' ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict if encoder_only: snake_case_ : Any = torch.load(__lowerCAmelCase , map_location=torch.device('''cpu''' ) ) else: snake_case_ : int = torch.load(__lowerCAmelCase , map_location=torch.device('''cpu''' ) )['''state_dict'''] # rename keys snake_case_ : Tuple = rename_keys(__lowerCAmelCase , encoder_only=__lowerCAmelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(__lowerCAmelCase , __lowerCAmelCase ) # create HuggingFace model and load state dict if encoder_only: snake_case_ : List[Any] = False snake_case_ : Union[str, Any] = SegformerForImageClassification(__lowerCAmelCase ) else: snake_case_ : Optional[Any] = SegformerForSemanticSegmentation(__lowerCAmelCase ) model.load_state_dict(__lowerCAmelCase ) model.eval() # forward pass snake_case_ : int = model(__lowerCAmelCase ) snake_case_ : Any = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": snake_case_ : int = torch.tensor( [ [[-4.6_310, -5.5_232, -6.2_356], [-5.1_921, -6.1_444, -6.5_996], [-5.4_424, -6.2_790, -6.7_574]], [[-12.1_391, -13.3_122, -13.9_554], [-12.8_732, -13.9_352, -14.3_563], [-12.9_438, -13.8_226, -14.2_513]], [[-12.5_134, -13.4_686, -14.4_915], [-12.8_669, -14.4_343, -14.7_758], [-13.2_523, -14.5_819, -15.0_694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": snake_case_ : str = torch.tensor( [ [[-7.5_820, -8.7_231, -8.3_215], [-8.0_600, -10.3_529, -10.0_304], [-7.5_208, -9.4_103, -9.6_239]], [[-12.6_918, -13.8_994, -13.7_137], [-13.3_196, -15.7_523, -15.4_789], [-12.9_343, -14.8_757, -14.9_689]], [[-11.1_911, -11.9_421, -11.3_243], [-11.3_342, -13.6_839, -13.3_581], [-10.3_909, -12.1_832, -12.4_858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": snake_case_ : str = torch.tensor( [ [[-11.8_173, -14.3_850, -16.3_128], [-14.5_648, -16.5_804, -18.6_568], [-14.7_223, -15.7_387, -18.4_218]], [[-15.7_290, -17.9_171, -19.4_423], [-18.3_105, -19.9_448, -21.4_661], [-17.9_296, -18.6_497, -20.7_910]], [[-15.0_783, -17.0_336, -18.2_789], [-16.8_771, -18.6_870, -20.1_612], [-16.2_454, -17.1_426, -19.5_055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": snake_case_ : List[str] = torch.tensor( [ [[-9.0_878, -10.2_081, -10.1_891], [-9.3_144, -10.7_941, -10.9_843], [-9.2_294, -10.3_855, -10.5_704]], [[-12.2_316, -13.9_068, -13.6_102], [-12.9_161, -14.3_702, -14.3_235], [-12.5_233, -13.7_174, -13.7_932]], [[-14.6_275, -15.2_490, -14.9_727], [-14.3_400, -15.9_687, -16.2_827], [-14.1_484, -15.4_033, -15.8_937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": snake_case_ : Dict = torch.tensor( [ [[-12.3_144, -13.2_447, -14.0_802], [-13.3_614, -14.5_816, -15.6_117], [-13.3_340, -14.4_433, -16.2_219]], [[-19.2_781, -20.4_128, -20.7_506], [-20.6_153, -21.6_566, -22.0_998], [-19.9_800, -21.0_430, -22.1_494]], [[-18.8_739, -19.7_804, -21.1_834], [-20.1_233, -21.6_765, -23.2_944], [-20.0_315, -21.2_641, -23.6_944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": snake_case_ : List[Any] = torch.tensor( [ [[-9.5_524, -12.0_835, -11.7_348], [-10.5_229, -13.6_446, -14.5_662], [-9.5_842, -12.8_851, -13.9_414]], [[-15.3_432, -17.5_323, -17.0_818], [-16.3_330, -18.9_255, -19.2_101], [-15.1_340, -17.7_848, -18.3_971]], [[-12.6_072, -14.9_486, -14.6_631], [-13.7_629, -17.0_907, -17.7_745], [-12.7_899, -16.1_695, -17.1_671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": snake_case_ : List[Any] = torch.tensor( [ [[-11.9_295, -13.4_057, -14.8_106], [-13.3_431, -14.8_179, -15.3_781], [-14.2_836, -15.5_942, -16.1_588]], [[-11.4_906, -12.8_067, -13.6_564], [-13.1_189, -14.0_500, -14.1_543], [-13.8_748, -14.5_136, -14.8_789]], [[0.5_374, 0.1_067, -0.4_742], [0.1_141, -0.2_255, -0.7_099], [-0.3_000, -0.5_924, -1.3_105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": snake_case_ : Optional[Any] = torch.tensor( [ [[-7.8_217, -9.8_767, -10.1_717], [-9.4_438, -10.9_058, -11.4_047], [-9.7_939, -12.3_495, -12.1_079]], [[-7.1_514, -9.5_336, -10.0_860], [-9.7_776, -11.6_822, -11.8_439], [-10.1_411, -12.7_655, -12.8_972]], [[0.3_021, 0.0_805, -0.2_310], [-0.0_328, -0.1_605, -0.2_714], [-0.1_408, -0.5_477, -0.6_976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": snake_case_ : Any = torch.tensor( [ [ [-1.1372E01, -1.2787E01, -1.3477E01], [-1.2536E01, -1.4194E01, -1.4409E01], [-1.3217E01, -1.4888E01, -1.5327E01], ], [ [-1.4791E01, -1.7122E01, -1.8277E01], [-1.7163E01, -1.9192E01, -1.9533E01], [-1.7897E01, -1.9991E01, -2.0315E01], ], [ [7.6723E-01, 4.1921E-01, -7.7878E-02], [4.7772E-01, 9.5557E-03, -2.8082E-01], [3.6032E-01, -2.4826E-01, -5.1168E-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": snake_case_ : Any = torch.tensor( [ [[-9.4_959, -11.3_087, -11.7_479], [-11.0_025, -12.6_540, -12.3_319], [-11.4_064, -13.0_487, -12.9_905]], [[-9.8_905, -11.3_084, -12.0_854], [-11.1_726, -12.7_698, -12.9_583], [-11.5_985, -13.3_278, -14.1_774]], [[0.2_213, 0.0_192, -0.2_466], [-0.1_731, -0.4_213, -0.4_874], [-0.3_126, -0.6_541, -1.1_389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": snake_case_ : Tuple = torch.tensor( [ [[-13.5_748, -13.9_111, -12.6_500], [-14.3_500, -15.3_683, -14.2_328], [-14.7_532, -16.0_424, -15.6_087]], [[-17.1_651, -15.8_725, -12.9_653], [-17.2_580, -17.3_718, -14.8_223], [-16.6_058, -16.8_783, -16.7_452]], [[-3.6_456, -3.0_209, -1.4_203], [-3.0_797, -3.1_959, -2.0_000], [-1.8_757, -1.9_217, -1.6_997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": snake_case_ : List[Any] = torch.tensor( [ [[-16.0_976, -16.4_856, -17.3_962], [-16.6_234, -19.0_342, -19.7_685], [-16.0_900, -18.0_661, -19.1_180]], [[-18.4_750, -18.8_488, -19.5_074], [-19.4_030, -22.1_570, -22.5_977], [-19.1_191, -20.8_486, -22.3_783]], [[-4.5_178, -5.5_037, -6.5_109], [-5.0_884, -7.2_174, -8.0_334], [-4.4_156, -5.8_117, -7.2_970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": snake_case_ : Optional[int] = torch.tensor( [ [[-14.2_081, -14.4_732, -14.1_977], [-14.5_867, -16.4_423, -16.6_356], [-13.4_441, -14.9_685, -16.8_696]], [[-14.4_576, -14.7_073, -15.0_451], [-15.0_816, -17.6_237, -17.9_873], [-14.4_213, -16.0_199, -18.5_992]], [[-4.7_349, -4.9_588, -5.0_966], [-4.3_210, -6.9_325, -7.2_591], [-3.4_312, -4.7_484, -7.1_917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": snake_case_ : str = torch.tensor( [ [[-11.7_737, -11.9_526, -11.3_273], [-13.6_692, -14.4_574, -13.8_878], [-13.8_937, -14.6_924, -15.9_345]], [[-14.6_706, -14.5_330, -14.1_306], [-16.1_502, -16.8_180, -16.4_269], [-16.8_338, -17.8_939, -20.1_746]], [[1.0_491, 0.8_289, 1.0_310], [1.1_044, 0.5_219, 0.8_055], [1.0_899, 0.6_926, 0.5_590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": snake_case_ : List[Any] = torch.tensor( [ [[-12.5_641, -13.4_777, -13.0_684], [-13.9_587, -15.8_983, -16.6_557], [-13.3_109, -15.7_350, -16.3_141]], [[-14.7_074, -15.4_352, -14.5_944], [-16.6_353, -18.1_663, -18.6_120], [-15.1_702, -18.0_329, -18.1_547]], [[-1.7_990, -2.0_951, -1.7_784], [-2.6_397, -3.8_245, -3.9_686], [-1.5_264, -2.8_126, -2.9_316]], ] ) else: snake_case_ : List[Any] = logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , __lowerCAmelCase , atol=1E-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) model.save_pretrained(__lowerCAmelCase ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) lowerCAmelCase_ = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
279
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase , cache_dir=UpperCAmelCase ) _UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(UpperCAmelCase , os.listdir(UpperCAmelCase )[0] , 'snapshots' ) )] _UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 4 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1e-3 assert np.abs(np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5e-1 _UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(UpperCAmelCase ) == num_samples def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , set_alpha_to_one=UpperCAmelCase , steps_offset=1 , ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=UpperCAmelCase , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = scheduler.create_state() _UpperCAmelCase = scheduler_state _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # With memory efficient attention _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , use_memory_efficient_attention=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
39
0
from collections import Counter from timeit import timeit def __UpperCamelCase ( lowerCAmelCase__ : List[str] = "" , ): return sum(c % 2 for c in Counter(input_str.replace(''' ''' , '''''' ).lower() ).values() ) < 2 def __UpperCamelCase ( lowerCAmelCase__ : List[str] = "" ): if len(__lowerCAmelCase ) == 0: return True __a : Dict = input_str.replace(''' ''' , '''''' ).lower() # character_freq_dict: Stores the frequency of every character in the input string __a : Tuple = {} for character in lower_case_input_str: __a : Tuple = character_freq_dict.get(__lowerCAmelCase , 0 ) + 1 __a : List[str] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def __UpperCamelCase ( lowerCAmelCase__ : Union[str, Any] = "" ): print('''\nFor string = ''' , __lowerCAmelCase , ''':''' ) print( '''> can_string_be_rearranged_as_palindrome_counter()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome_counter(__lowerCAmelCase ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome_counter(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) print( '''> can_string_be_rearranged_as_palindrome()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome(__lowerCAmelCase ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) if __name__ == "__main__": lowercase__ =input( 'Enter string to determine if it can be rearranged as a palindrome or not: ' ).strip() benchmark(check_str) lowercase__ =can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {"" if status else "not "}be rearranged as a palindrome""")
216
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = AlbertTokenizer UpperCamelCase__ = AlbertTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = True def UpperCamelCase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 'this is a test' _UpperCAmelCase = 'this is a test' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = '<pad>' _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(UpperCAmelCase ) , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = 'I was born in 92000, and this is falsé.' _UpperCAmelCase = tokenizer.tokenize(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [48, 25, 21, 1289] ) _UpperCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) _UpperCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode('sequence builders' ) _UpperCAmelCase = tokenizer.encode('multi-sequence build' ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
39
0
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class __magic_name__ : def __init__( self , __snake_case , __snake_case=13 , __snake_case=7 , __snake_case=True , __snake_case=True , __snake_case=99 , __snake_case=32 , __snake_case=5 , __snake_case=4 , __snake_case=37 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=50 , __snake_case=0.02 , __snake_case=True , __snake_case=None , ) -> Tuple: '''simple docstring''' __a =parent __a =batch_size __a =seq_length __a =is_training __a =use_input_mask __a =vocab_size __a =hidden_size __a =num_hidden_layers __a =num_attention_heads __a =intermediate_size __a =hidden_act __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =max_position_embeddings __a =initializer_range __a =use_labels __a =scope def __magic_name__ ( self ) -> str: '''simple docstring''' __a =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a =None if self.use_input_mask: __a =random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: __a =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a =self.get_config() return config, input_ids, input_mask, token_labels def __magic_name__ ( self ) -> int: '''simple docstring''' return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def __magic_name__ ( self ) -> List[str]: '''simple docstring''' ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) =self.prepare_config_and_inputs() __a =True __a =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case , __snake_case , **__snake_case , ) -> int: '''simple docstring''' __a =BertGenerationEncoder(config=__snake_case ) model.to(__snake_case ) model.eval() __a =model(__snake_case , attention_mask=__snake_case ) __a =model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , **__snake_case , ) -> Union[str, Any]: '''simple docstring''' __a =True __a =BertGenerationEncoder(config=__snake_case ) model.to(__snake_case ) model.eval() __a =model( __snake_case , attention_mask=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) __a =model( __snake_case , attention_mask=__snake_case , encoder_hidden_states=__snake_case , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , **__snake_case , ) -> int: '''simple docstring''' __a =True __a =True __a =BertGenerationDecoder(config=__snake_case ).to(__snake_case ).eval() # first forward pass __a =model( __snake_case , attention_mask=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , use_cache=__snake_case , ) __a =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __a =ids_tensor((self.batch_size, 3) , config.vocab_size ) __a =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __a =torch.cat([input_ids, next_tokens] , dim=-1 ) __a =torch.cat([input_mask, next_mask] , dim=-1 ) __a =model( __snake_case , attention_mask=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , output_hidden_states=__snake_case , )['hidden_states'][0] __a =model( __snake_case , attention_mask=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , past_key_values=__snake_case , output_hidden_states=__snake_case , )['hidden_states'][0] # select random slice __a =ids_tensor((1,) , output_from_past.shape[-1] ).item() __a =output_from_no_past[:, -3:, random_slice_idx].detach() __a =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__snake_case , __snake_case , atol=1e-3 ) ) def __magic_name__ ( self , __snake_case , __snake_case , __snake_case , __snake_case , *__snake_case , ) -> Any: '''simple docstring''' __a =BertGenerationDecoder(__snake_case ) model.to(__snake_case ) model.eval() __a =model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __magic_name__ ( self ) -> str: '''simple docstring''' __a , __a , __a , __a =self.prepare_config_and_inputs() __a ={'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __magic_name__ ( snake_case__ , snake_case__ , snake_case__ , unittest.TestCase ): SCREAMING_SNAKE_CASE = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () SCREAMING_SNAKE_CASE = (BertGenerationDecoder,) if is_torch_available() else () SCREAMING_SNAKE_CASE = ( {'feature-extraction': BertGenerationEncoder, 'text-generation': BertGenerationDecoder} if is_torch_available() else {} ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =BertGenerationEncoderTester(self ) __a =ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' __a , __a , __a , __a =self.model_tester.prepare_config_and_inputs() __a ='bert' self.model_tester.create_and_check_model(__snake_case , __snake_case , __snake_case , __snake_case ) def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__snake_case ) def __magic_name__ ( self ) -> str: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__snake_case ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) =self.model_tester.prepare_config_and_inputs_for_decoder() __a =None self.model_tester.create_and_check_model_as_decoder( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , ) def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*__snake_case ) @slow def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' __a =BertGenerationEncoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) self.assertIsNotNone(__snake_case ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> str: '''simple docstring''' __a =BertGenerationEncoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) __a =torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 1_0140, 102]] ) with torch.no_grad(): __a =model(__snake_case )[0] __a =torch.Size([1, 8, 1024] ) self.assertEqual(output.shape , __snake_case ) __a =torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1e-4 ) ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a =BertGenerationDecoder.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) __a =torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 1_0140, 102]] ) with torch.no_grad(): __a =model(__snake_case )[0] __a =torch.Size([1, 8, 5_0358] ) self.assertEqual(output.shape , __snake_case ) __a =torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1e-4 ) )
218
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def _UpperCamelCase ( ) -> Tuple: import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join __UpperCAmelCase : str = "__test_patch_submodule_mock__" with patch_submodule(_test_patching, "os.path.join", __lowerCAmelCase ): # Every way to access os.path.join must be patched, and the rest must stay untouched # check os.path.join assert isinstance(_test_patching.os, _PatchedModuleObj ) assert isinstance(_test_patching.os.path, _PatchedModuleObj ) assert _test_patching.os.path.join is mock # check path.join assert isinstance(_test_patching.path, _PatchedModuleObj ) assert _test_patching.path.join is mock # check join assert _test_patching.join is mock # check that the other attributes are untouched assert _test_patching.os.rename is original_rename assert _test_patching.path.dirname is original_dirname assert _test_patching.os.path.dirname is original_dirname # Even renamed modules or objects must be patched # check renamed_os.path.join assert isinstance(_test_patching.renamed_os, _PatchedModuleObj ) assert isinstance(_test_patching.renamed_os.path, _PatchedModuleObj ) assert _test_patching.renamed_os.path.join is mock # check renamed_path.join assert isinstance(_test_patching.renamed_path, _PatchedModuleObj ) assert _test_patching.renamed_path.join is mock # check renamed_join assert _test_patching.renamed_join is mock # check that the other attributes are untouched assert _test_patching.renamed_os.rename is original_rename assert _test_patching.renamed_path.dirname is original_dirname assert _test_patching.renamed_os.path.dirname is original_dirname # check that everthing is back to normal when the patch is over assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join def _UpperCamelCase ( ) -> List[Any]: assert _test_patching.open is open __UpperCAmelCase : str = "__test_patch_submodule_builtin_mock__" # _test_patching has "open" in its globals assert _test_patching.open is open with patch_submodule(_test_patching, "open", __lowerCAmelCase ): assert _test_patching.open is mock # check that everthing is back to normal when the patch is over assert _test_patching.open is open def _UpperCamelCase ( ) -> int: __UpperCAmelCase : str = "__test_patch_submodule_missing_mock__" with patch_submodule(_test_patching, "pandas.read_csv", __lowerCAmelCase ): pass def _UpperCamelCase ( ) -> Optional[Any]: __UpperCAmelCase : Union[str, Any] = "__test_patch_submodule_missing_builtin_mock__" # _test_patching doesn't have "len" in its globals assert getattr(_test_patching, "len", __lowerCAmelCase ) is None with patch_submodule(_test_patching, "len", __lowerCAmelCase ): assert _test_patching.len is mock assert _test_patching.len is len def _UpperCamelCase ( ) -> Optional[Any]: __UpperCAmelCase : int = "__test_patch_submodule_start_and_stop_mock__" __UpperCAmelCase : List[str] = patch_submodule(_test_patching, "open", __lowerCAmelCase ) assert _test_patching.open is open patch.start() assert _test_patching.open is mock patch.stop() assert _test_patching.open is open def _UpperCamelCase ( ) -> Optional[int]: from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join __UpperCAmelCase : List[Any] = "__test_patch_submodule_successive_join__" __UpperCAmelCase : List[str] = "__test_patch_submodule_successive_dirname__" __UpperCAmelCase : int = "__test_patch_submodule_successive_rename__" assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename with patch_submodule(_test_patching, "os.path.join", __lowerCAmelCase ): with patch_submodule(_test_patching, "os.rename", __lowerCAmelCase ): with patch_submodule(_test_patching, "os.path.dirname", __lowerCAmelCase ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename # try another order with patch_submodule(_test_patching, "os.rename", __lowerCAmelCase ): with patch_submodule(_test_patching, "os.path.join", __lowerCAmelCase ): with patch_submodule(_test_patching, "os.path.dirname", __lowerCAmelCase ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename def _UpperCamelCase ( ) -> Any: __UpperCAmelCase : int = "__test_patch_submodule_doesnt_exist_mock__" with patch_submodule(_test_patching, "__module_that_doesn_exist__.__attribute_that_doesn_exist__", __lowerCAmelCase ): pass with patch_submodule(_test_patching, "os.__attribute_that_doesn_exist__", __lowerCAmelCase ): pass
157
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "distilbert" UpperCamelCase__ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , UpperCAmelCase=3_0522 , UpperCAmelCase=512 , UpperCAmelCase=False , UpperCAmelCase=6 , UpperCAmelCase=12 , UpperCAmelCase=768 , UpperCAmelCase=4 * 768 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=0.02 , UpperCAmelCase=0.1 , UpperCAmelCase=0.2 , UpperCAmelCase=0 , **UpperCAmelCase , ): """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = sinusoidal_pos_embds _UpperCAmelCase = n_layers _UpperCAmelCase = n_heads _UpperCAmelCase = dim _UpperCAmelCase = hidden_dim _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation _UpperCAmelCase = initializer_range _UpperCAmelCase = qa_dropout _UpperCAmelCase = seq_classif_dropout super().__init__(**UpperCAmelCase , pad_token_id=UpperCAmelCase ) class __lowerCamelCase ( snake_case__): """simple docstring""" @property def UpperCamelCase ( self ): """simple docstring""" if self.task == "multiple-choice": _UpperCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _UpperCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
39
0
'''simple docstring''' from __future__ import annotations from fractions import Fraction def __a(SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[int] ): '''simple docstring''' return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __a(SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' _lowerCAmelCase = [] _lowerCAmelCase = 11 _lowerCAmelCase = int("1" + "0" * digit_len ) for num in range(__lowerCAmelCase , __lowerCAmelCase ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(__lowerCAmelCase , __lowerCAmelCase ): solutions.append(F'''{num}/{den}''' ) den += 1 num += 1 _lowerCAmelCase = 10 return solutions def __a(SCREAMING_SNAKE_CASE_ : Tuple = 2 ): '''simple docstring''' _lowerCAmelCase = 1.0 for fraction in fraction_list(__lowerCAmelCase ): _lowerCAmelCase = Fraction(__lowerCAmelCase ) result *= frac.denominator / frac.numerator return int(__lowerCAmelCase ) if __name__ == "__main__": print(solution())
158
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() _a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCamelCase ( snake_case__): """simple docstring""" def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {'source': 'What is love ?', 'target': 'life'} _UpperCAmelCase = {'train': 12, 'val': 2, 'test': 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: _UpperCAmelCase = '\n'.join([contents[field]] * n_lines[split] ) with open(os.path.join(UpperCAmelCase , F"""{split}.{field}""" ) , 'w' ) as f: f.write(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase = "pytorch" ): """simple docstring""" _UpperCAmelCase = self.get_auto_remove_tmp_dir() _UpperCAmelCase = os.path.join(UpperCAmelCase , 'output' ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'data' ) self._create_dummy_data(data_dir=UpperCAmelCase ) _UpperCAmelCase = F""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(F"""--gpus={gpus}""" ) if is_apex_available(): testargs.append('--fp16' ) else: testargs.append('--gpus=0' ) testargs.append('--distributed_backend=ddp_cpu' ) testargs.append('--num_processes=2' ) _UpperCAmelCase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(UpperCAmelCase , env=self.get_env() ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'metrics.json' ) with open(UpperCAmelCase ) as f: _UpperCAmelCase = json.load(UpperCAmelCase ) return result @require_torch_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 )
39
0
from functools import reduce lowerCamelCase__ = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = N ) -> int: return max( # mypy cannot properly interpret reduce int(reduce(lambda SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str(int(__lowerCAmelCase ) * int(__lowerCAmelCase ) ) , n[i : i + 13] ) ) for i in range(len(__lowerCAmelCase ) - 12 ) ) if __name__ == "__main__": print(F"""{solution() = }""")
212
class __lowerCamelCase : """simple docstring""" def __init__( self ): """simple docstring""" _UpperCAmelCase = {} # Mapping from char to TrieNode _UpperCAmelCase = False def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: _UpperCAmelCase = TrieNode() _UpperCAmelCase = curr.nodes[char] _UpperCAmelCase = True def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: return False _UpperCAmelCase = curr.nodes[char] return curr.is_leaf def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" def _delete(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: if index == len(UpperCAmelCase ): # If word does not exist if not curr.is_leaf: return False _UpperCAmelCase = False return len(curr.nodes ) == 0 _UpperCAmelCase = word[index] _UpperCAmelCase = curr.nodes.get(UpperCAmelCase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted _UpperCAmelCase = _delete(UpperCAmelCase , UpperCAmelCase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCAmelCase , 0 ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" if node.is_leaf: print(__lowerCAmelCase , end=' ' ) for key, value in node.nodes.items(): print_words(__lowerCAmelCase , word + key ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = TrieNode() root.insert_many(__lowerCAmelCase ) # print_words(root, "") assert all(root.find(__lowerCAmelCase ) for word in words ) assert root.find('banana' ) assert not root.find('bandanas' ) assert not root.find('apps' ) assert root.find('apple' ) assert root.find('all' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" print(str(__lowerCAmelCase ) , 'works!' if passes else 'doesn\'t work :(' ) def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" print_results('Testing trie functionality' , test_trie() ) if __name__ == "__main__": main()
39
0
from __future__ import annotations def UpperCAmelCase ( a_ ) -> list[int]: """simple docstring""" __A = 2 __A = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCAmelCase ) if n > 1: factors.append(__lowerCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
15
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _a = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class __lowerCamelCase ( unittest.TestCase): """simple docstring""" UpperCamelCase__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = ZeroShotClassificationPipeline( model=UpperCAmelCase , tokenizer=UpperCAmelCase , candidate_labels=['polics', 'health'] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # No kwarg _UpperCAmelCase = classifier('Who are you voting for in 2020?' , ['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics, public health' ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics', 'public health'] ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='This text is about {}' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # https://github.com/huggingface/transformers/issues/13846 _UpperCAmelCase = classifier(['I am happy'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(1 ) ] , ) _UpperCAmelCase = classifier(['I am happy', 'I am sad'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(2 ) ] , ) with self.assertRaises(UpperCAmelCase ): classifier('' , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier(UpperCAmelCase , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels='' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='Not formatting template' , ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template=UpperCAmelCase , ) self.run_entailment_id(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = zero_shot_classifier.model.config _UpperCAmelCase = config.labelaid _UpperCAmelCase = zero_shot_classifier.entailment_id _UpperCAmelCase = {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) _UpperCAmelCase = {'entailment': 0, 'neutral': 1, 'contradiction': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 0, 'NON-ENTAIL': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 2, 'NEUTRAL': 1, 'CONTR': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) _UpperCAmelCase = original_labelaid self.assertEqual(UpperCAmelCase , zero_shot_classifier.entailment_id ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( 'Who are you voting for in 2020?' * 100 , candidate_labels=['politics', 'public health', 'science'] ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='tf' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @slow @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='pt' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , ) @slow @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='tf' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , )
39
0
from collections import deque from math import floor from random import random from time import time class __snake_case : def __init__( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = {} def __a ( self : Optional[Any] , _lowercase : List[str] , _lowercase : int , _lowercase : Optional[int]=1 ): """simple docstring""" if self.graph.get(_lowercase ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: SCREAMING_SNAKE_CASE__ = [[w, v]] if not self.graph.get(_lowercase ): SCREAMING_SNAKE_CASE__ = [] def __a ( self : Optional[Any] ): """simple docstring""" return list(self.graph ) def __a ( self : int , _lowercase : Any , _lowercase : Dict ): """simple docstring""" if self.graph.get(_lowercase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowercase ) def __a ( self : Tuple , _lowercase : Any=-2 , _lowercase : int=-1 ): """simple docstring""" if s == d: return [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] if s == -2: SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowercase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return visited def __a ( self : Any , _lowercase : str=-1 ): """simple docstring""" if c == -1: SCREAMING_SNAKE_CASE__ = floor(random() * 1_00_00 ) + 10 for i in range(_lowercase ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): SCREAMING_SNAKE_CASE__ = floor(random() * c ) + 1 if n != i: self.add_pair(_lowercase , _lowercase , 1 ) def __a ( self : str , _lowercase : Dict=-2 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = deque() SCREAMING_SNAKE_CASE__ = [] if s == -2: SCREAMING_SNAKE_CASE__ = list(self.graph )[0] d.append(_lowercase ) visited.append(_lowercase ) while d: SCREAMING_SNAKE_CASE__ = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __a ( self : Optional[Any] , _lowercase : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def __a ( self : str , _lowercase : int ): """simple docstring""" return len(self.graph[u] ) def __a ( self : Tuple , _lowercase : List[str]=-2 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] if s == -2: SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return sorted_nodes def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = -2 SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ = len(_lowercase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ = True if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = False indirect_parents.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return list(_lowercase ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = -2 SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ = len(_lowercase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ = True if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = False indirect_parents.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return False def __a ( self : Dict , _lowercase : Optional[int]=-2 , _lowercase : Union[str, Any]=-1 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = time() self.dfs(_lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = time() return end - begin def __a ( self : str , _lowercase : str=-2 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = time() self.bfs(_lowercase ) SCREAMING_SNAKE_CASE__ = time() return end - begin class __snake_case : def __init__( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = {} def __a ( self : List[str] , _lowercase : int , _lowercase : str , _lowercase : Optional[Any]=1 ): """simple docstring""" if self.graph.get(_lowercase ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist SCREAMING_SNAKE_CASE__ = [[w, v]] # add the other way if self.graph.get(_lowercase ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist SCREAMING_SNAKE_CASE__ = [[w, u]] def __a ( self : Optional[Any] , _lowercase : Optional[Any] , _lowercase : Dict ): """simple docstring""" if self.graph.get(_lowercase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowercase ) # the other way round if self.graph.get(_lowercase ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(_lowercase ) def __a ( self : Any , _lowercase : List[str]=-2 , _lowercase : Any=-1 ): """simple docstring""" if s == d: return [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] if s == -2: SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowercase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return visited def __a ( self : str , _lowercase : Union[str, Any]=-1 ): """simple docstring""" if c == -1: SCREAMING_SNAKE_CASE__ = floor(random() * 1_00_00 ) + 10 for i in range(_lowercase ): # every vertex has max 100 edges for _ in range(floor(random() * 1_02 ) + 1 ): SCREAMING_SNAKE_CASE__ = floor(random() * c ) + 1 if n != i: self.add_pair(_lowercase , _lowercase , 1 ) def __a ( self : Optional[Any] , _lowercase : Union[str, Any]=-2 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = deque() SCREAMING_SNAKE_CASE__ = [] if s == -2: SCREAMING_SNAKE_CASE__ = list(self.graph )[0] d.append(_lowercase ) visited.append(_lowercase ) while d: SCREAMING_SNAKE_CASE__ = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __a ( self : int , _lowercase : List[str] ): """simple docstring""" return len(self.graph[u] ) def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = -2 SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ = len(_lowercase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ = True if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = False indirect_parents.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return list(_lowercase ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = list(self.graph )[0] stack.append(_lowercase ) visited.append(_lowercase ) SCREAMING_SNAKE_CASE__ = -2 SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: SCREAMING_SNAKE_CASE__ = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): SCREAMING_SNAKE_CASE__ = len(_lowercase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) SCREAMING_SNAKE_CASE__ = node[1] break # check if all the children are visited if s == ss: stack.pop() SCREAMING_SNAKE_CASE__ = True if len(_lowercase ) != 0: SCREAMING_SNAKE_CASE__ = stack[len(_lowercase ) - 1] else: SCREAMING_SNAKE_CASE__ = False indirect_parents.append(_lowercase ) SCREAMING_SNAKE_CASE__ = s SCREAMING_SNAKE_CASE__ = ss # check if se have reached the starting point if len(_lowercase ) == 0: return False def __a ( self : List[str] ): """simple docstring""" return list(self.graph ) def __a ( self : List[str] , _lowercase : Optional[int]=-2 , _lowercase : int=-1 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = time() self.dfs(_lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = time() return end - begin def __a ( self : Union[str, Any] , _lowercase : Any=-2 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = time() self.bfs(_lowercase ) SCREAMING_SNAKE_CASE__ = time() return end - begin
219
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _a = get_logger(__name__) class __lowerCamelCase ( enum.Enum): """simple docstring""" UpperCamelCase__ = "all_checks" UpperCamelCase__ = "basic_checks" UpperCamelCase__ = "no_checks" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None )-> str: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase = ' for ' + verification_name if verification_name is not None else '' if len(__lowerCAmelCase ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__lowerCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__lowerCAmelCase ) ) logger.info('All the splits matched successfully.' ) def __A ( __lowerCAmelCase , __lowerCAmelCase = True )-> dict: """simple docstring""" if record_checksum: _UpperCAmelCase = shaaaa() with open(__lowerCAmelCase , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__lowerCAmelCase ) _UpperCAmelCase = m.hexdigest() else: _UpperCAmelCase = None return {"num_bytes": os.path.getsize(__lowerCAmelCase ), "checksum": checksum} def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
39
0
from collections.abc import Generator from math import sin def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Tuple ) -> bytes: '''simple docstring''' if len(__lowerCAmelCase ) != 3_2: raise ValueError("Input must be of length 32" ) A__ = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Tuple ) -> bytes: '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) A__ = format(__lowerCAmelCase , "08x" )[-8:] A__ = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Any ) -> bytes: '''simple docstring''' A__ = b"" for char in message: bit_string += format(__lowerCAmelCase , "08b" ).encode("utf-8" ) A__ = format(len(__lowerCAmelCase ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__lowerCAmelCase ) % 5_1_2 != 4_4_8: bit_string += b"0" bit_string += to_little_endian(start_len[3_2:] ) + to_little_endian(start_len[:3_2] ) return bit_string def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Tuple ) -> Generator[list[int], None, None]: '''simple docstring''' if len(__lowerCAmelCase ) % 5_1_2 != 0: raise ValueError("Input must have length that\'s a multiple of 512" ) for pos in range(0 , len(__lowerCAmelCase ) , 5_1_2 ): A__ = bit_string[pos : pos + 5_1_2] A__ = [] for i in range(0 , 5_1_2 , 3_2 ): block_words.append(int(to_little_endian(block[i : i + 3_2] ) , 2 ) ) yield block_words def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Optional[Any] ) -> int: '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) A__ = format(__lowerCAmelCase , "032b" ) A__ = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(__lowerCAmelCase , 2 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: str ) -> int: '''simple docstring''' return (a + b) % 2**3_2 def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Union[str, Any] , SCREAMING_SNAKE_CASE_: Optional[int] ) -> int: '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (3_2 - shift))) % 2**3_2 def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Union[str, Any] ) -> bytes: '''simple docstring''' A__ = preprocess(__lowerCAmelCase ) A__ = [int(2**3_2 * abs(sin(i + 1 ) ) ) for i in range(6_4 )] # Starting states A__ = 0x67_45_23_01 A__ = 0xef_cd_ab_89 A__ = 0x98_ba_dc_fe A__ = 0x10_32_54_76 A__ = [ 7, 1_2, 1_7, 2_2, 7, 1_2, 1_7, 2_2, 7, 1_2, 1_7, 2_2, 7, 1_2, 1_7, 2_2, 5, 9, 1_4, 2_0, 5, 9, 1_4, 2_0, 5, 9, 1_4, 2_0, 5, 9, 1_4, 2_0, 4, 1_1, 1_6, 2_3, 4, 1_1, 1_6, 2_3, 4, 1_1, 1_6, 2_3, 4, 1_1, 1_6, 2_3, 6, 1_0, 1_5, 2_1, 6, 1_0, 1_5, 2_1, 6, 1_0, 1_5, 2_1, 6, 1_0, 1_5, 2_1, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__lowerCAmelCase ): A__ = aa A__ = ba A__ = ca A__ = da # Hash current chunk for i in range(6_4 ): if i <= 1_5: # f = (b & c) | (not_32(b) & d) # Alternate definition for f A__ = d ^ (b & (c ^ d)) A__ = i elif i <= 3_1: # f = (d & b) | (not_32(d) & c) # Alternate definition for f A__ = c ^ (d & (b ^ c)) A__ = (5 * i + 1) % 1_6 elif i <= 4_7: A__ = b ^ c ^ d A__ = (3 * i + 5) % 1_6 else: A__ = c ^ (b | not_aa(__lowerCAmelCase )) A__ = (7 * i) % 1_6 A__ = (f + a + added_consts[i] + block_words[g]) % 2**3_2 A__ = d A__ = c A__ = b A__ = sum_aa(__lowerCAmelCase , left_rotate_aa(__lowerCAmelCase , shift_amounts[i] ) ) # Add hashed chunk to running total A__ = sum_aa(__lowerCAmelCase , __lowerCAmelCase ) A__ = sum_aa(__lowerCAmelCase , __lowerCAmelCase ) A__ = sum_aa(__lowerCAmelCase , __lowerCAmelCase ) A__ = sum_aa(__lowerCAmelCase , __lowerCAmelCase ) A__ = reformat_hex(__lowerCAmelCase ) + reformat_hex(__lowerCAmelCase ) + reformat_hex(__lowerCAmelCase ) + reformat_hex(__lowerCAmelCase ) return digest if __name__ == "__main__": import doctest doctest.testmod()
68
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=16 , UpperCAmelCase=[1, 2, 1] , UpperCAmelCase=[2, 2, 4] , UpperCAmelCase=2 , UpperCAmelCase=2.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=10 , UpperCAmelCase=8 , UpperCAmelCase=["stage1", "stage2", "stage3"] , UpperCAmelCase=[1, 2, 3] , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = embed_dim _UpperCAmelCase = depths _UpperCAmelCase = num_heads _UpperCAmelCase = window_size _UpperCAmelCase = mlp_ratio _UpperCAmelCase = qkv_bias _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = drop_path_rate _UpperCAmelCase = hidden_act _UpperCAmelCase = use_absolute_embeddings _UpperCAmelCase = patch_norm _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = initializer_range _UpperCAmelCase = is_training _UpperCAmelCase = scope _UpperCAmelCase = use_labels _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = encoder_stride _UpperCAmelCase = out_features _UpperCAmelCase = out_indices def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) _UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCAmelCase ): _UpperCAmelCase = ['stem'] _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCamelCase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) # Swin has a different seq_length _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = 3 _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCAmelCase ): _UpperCAmelCase = 0 return t def check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase={} ): with torch.no_grad(): _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ).to_tuple() def recursive_check(UpperCAmelCase , UpperCAmelCase ): if isinstance(UpperCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase , UpperCAmelCase ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCAmelCase ) , set_nan_tensor_to_zero(UpperCAmelCase ) , atol=1e-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}. Dict has""" F""" `nan`: {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}.""" ) , ) recursive_check(UpperCAmelCase , UpperCAmelCase ) for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) @require_torch class __lowerCamelCase ( unittest.TestCase , snake_case__): """simple docstring""" UpperCamelCase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCamelCase__ = MaskFormerSwinConfig def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: _UpperCAmelCase = backbone_class(UpperCAmelCase ) backbone.to(UpperCAmelCase ) backbone.eval() _UpperCAmelCase = backbone(**UpperCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _UpperCAmelCase = backbone(**UpperCAmelCase , output_hidden_states=UpperCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _UpperCAmelCase = backbone(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertIsNotNone(outputs.attentions )
39
0
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() UpperCAmelCase__ : Tuple = { 'bart': ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'bert': ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'bert-base-cased-finetuned-mrpc': ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'dpr': ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'gpt2': ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlnet': ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm': ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'xlm-roberta': ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'transfo-xl': ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'openai-gpt': ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'roberta': ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'layoutlm': ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), 'roberta-large-mnli': ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'camembert': ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'flaubert': ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert': ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'distilbert-base-distilled-squad': ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert': ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'lxmert-visual-feature-encoder': ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'ctrl': ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'albert': ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 't5': ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'electra': ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), 'wav2vec2': ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def lowerCamelCase__ ( a , a , a , a , a=False , a=True ) -> Optional[Any]: if model_type not in MODEL_CLASSES: raise ValueError(f"""Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.""" ) _A , _A , _A , _A: Union[str, Any] = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: _A: Union[str, Any] = cached_file(__lowerCAmelCase , __lowerCAmelCase , force_download=not use_cached_models ) _A: Optional[int] = config_class.from_json_file(__lowerCAmelCase ) _A: int = True _A: Optional[int] = True print(f"""Building TensorFlow model from configuration: {config}""" ) _A: Dict = model_class(__lowerCAmelCase ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): _A: List[Any] = cached_file( __lowerCAmelCase , __lowerCAmelCase , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: _A: List[str] = load_pytorch_checkpoint_in_tfa_model(__lowerCAmelCase , __lowerCAmelCase ) if compare_with_pt_model: _A: Tuple = tf_model(tf_model.dummy_inputs , training=__lowerCAmelCase ) # build the network _A: int = torch.load(__lowerCAmelCase , map_location='''cpu''' ) _A: Optional[Any] = pt_model_class.from_pretrained( pretrained_model_name_or_path=__lowerCAmelCase , config=__lowerCAmelCase , state_dict=__lowerCAmelCase ) with torch.no_grad(): _A: Any = pt_model(**pt_model.dummy_inputs ) _A: int = pto[0].numpy() _A: int = tfo[0].numpy() _A: Dict = np.amax(np.abs(np_pt - np_tf ) ) print(f"""Max absolute difference between models outputs {diff}""" ) assert diff <= 2E-2, f"""Error, model absolute difference is >2e-2: {diff}""" # Save pytorch-model print(f"""Save TensorFlow model to {tf_dump_path}""" ) tf_model.save_weights(__lowerCAmelCase , save_format='''h5''' ) def lowerCamelCase__ ( a , a , a=None , a=None , a=False , a=False , a=False , a=False , ) -> Tuple: if args_model_type is None: _A: List[Any] = list(MODEL_CLASSES.keys() ) else: _A: Optional[int] = [args_model_type] for j, model_type in enumerate(__lowerCAmelCase , start=1 ): print('''=''' * 1_00 ) print(f""" Converting model type {j}/{len(__lowerCAmelCase )}: {model_type}""" ) print('''=''' * 1_00 ) if model_type not in MODEL_CLASSES: raise ValueError(f"""Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.""" ) _A , _A , _A , _A , _A: Union[str, Any] = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: _A: Optional[Any] = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: _A: List[str] = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(__lowerCAmelCase , __lowerCAmelCase ) , start=1 ): print('''-''' * 1_00 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f""" Skipping finetuned checkpoint {model_shortcut_name}""" ) continue _A: Tuple = model_shortcut_name elif only_convert_finetuned_models: print(f""" Skipping not finetuned checkpoint {model_shortcut_name}""" ) continue print( f""" Converting checkpoint {i}/{len(__lowerCAmelCase )}: {model_shortcut_name} - model_type {model_type}""" ) print('''-''' * 1_00 ) if config_shortcut_name in aws_config_map: _A: Any = cached_file(__lowerCAmelCase , __lowerCAmelCase , force_download=not use_cached_models ) else: _A: List[str] = config_shortcut_name if model_shortcut_name in aws_model_maps: _A: Tuple = cached_file(__lowerCAmelCase , __lowerCAmelCase , force_download=not use_cached_models ) else: _A: Any = model_shortcut_name if os.path.isfile(__lowerCAmelCase ): _A: Optional[Any] = '''converted_model''' convert_pt_checkpoint_to_tf( model_type=__lowerCAmelCase , pytorch_checkpoint_path=__lowerCAmelCase , config_file=__lowerCAmelCase , tf_dump_path=os.path.join(__lowerCAmelCase , model_shortcut_name + '''-tf_model.h5''' ) , compare_with_pt_model=__lowerCAmelCase , ) if remove_cached_files: os.remove(__lowerCAmelCase ) os.remove(__lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_dump_path', default=None, type=str, required=True, help='Path to the output Tensorflow dump file.' ) parser.add_argument( '--model_type', default=None, type=str, help=( F"""Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and """ 'convert all the models from AWS.' ), ) parser.add_argument( '--pytorch_checkpoint_path', default=None, type=str, help=( 'Path to the PyTorch checkpoint path or shortcut name to download from AWS. ' 'If not given, will download and convert all the checkpoints from AWS.' ), ) parser.add_argument( '--config_file', default=None, type=str, help=( 'The config json file corresponding to the pre-trained model. \n' 'This specifies the model architecture. If not given and ' '--pytorch_checkpoint_path is not given or is a shortcut name ' 'use the configuration associated to the shortcut name on the AWS' ), ) parser.add_argument( '--compare_with_pt_model', action='store_true', help='Compare Tensorflow and PyTorch model predictions.' ) parser.add_argument( '--use_cached_models', action='store_true', help='Use cached models if possible instead of updating to latest checkpoint versions.', ) parser.add_argument( '--remove_cached_files', action='store_true', help='Remove pytorch models after conversion (save memory when converting in batches).', ) parser.add_argument('--only_convert_finetuned_models', action='store_true', help='Only convert finetuned models.') UpperCAmelCase__ : Dict = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
121
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = TransfoXLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" super().setUp() _UpperCAmelCase = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = '<unk> UNwanted , running' _UpperCAmelCase = '<unk> unwanted, running' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(UpperCAmelCase , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [0, 4, 8, 7] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) _UpperCAmelCase = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' _UpperCAmelCase = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(UpperCAmelCase ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(UpperCAmelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
39
0
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets _a = datasets.logging.get_logger(__name__) _a = """\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = \"{COMET}: A Neural Framework for {MT} Evaluation\", author = \"Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon\", booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\", month = nov, year = \"2020\", address = \"Online\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\", pages = \"2685--2702\", } """ _a = """\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. """ _a = """ COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric(\'comet\') >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"] >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"] >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results[\"scores\"]]) [0.19, 0.92] """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase( datasets.Metric ): def UpperCAmelCase ( self) -> int: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://unbabel.github.io/COMET/html/index.html''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''sources''': datasets.Value('''string''' , id='''sequence'''), '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , codebase_urls=['''https://github.com/Unbabel/COMET'''] , reference_urls=[ '''https://github.com/Unbabel/COMET''', '''https://www.aclweb.org/anthology/2020.emnlp-main.213/''', '''http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6''', ] , ) def UpperCAmelCase ( self , __a) -> List[Any]: '''simple docstring''' if self.config_name == "default": _UpperCamelCase = comet.load_from_checkpoint(comet.download_model('''wmt20-comet-da''')) else: _UpperCamelCase = comet.load_from_checkpoint(comet.download_model(self.config_name)) def UpperCAmelCase ( self , __a , __a , __a , __a=None , __a=False) -> Union[str, Any]: '''simple docstring''' if gpus is None: _UpperCamelCase = 1 if torch.cuda.is_available() else 0 _UpperCamelCase = {'''src''': sources, '''mt''': predictions, '''ref''': references} _UpperCamelCase = [dict(zip(__a , __a)) for t in zip(*data.values())] _UpperCamelCase , _UpperCamelCase = self.scorer.predict(__a , gpus=__a , progress_bar=__a) return {"mean_score": mean_score, "scores": scores}
194
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = {'''configuration_opt''': ['''OPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OPTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''OPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OPTForCausalLM''', '''OPTModel''', '''OPTPreTrainedModel''', '''OPTForSequenceClassification''', '''OPTForQuestionAnswering''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['''TFOPTForCausalLM''', '''TFOPTModel''', '''TFOPTPreTrainedModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''FlaxOPTForCausalLM''', '''FlaxOPTModel''', '''FlaxOPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
39
0
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = '''hf-internal-testing/tiny-random-t5''' snake_case_ : str = AutoTokenizer.from_pretrained(__magic_name__ ) snake_case_ : int = AutoModelForSeqaSeqLM.from_pretrained(__magic_name__ ) snake_case_ : Tuple = tokenizer('''This is me''' , return_tensors='''pt''' ) snake_case_ : Dict = model.to_bettertransformer() self.assertTrue(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) snake_case_ : List[str] = model.generate(**__magic_name__ ) snake_case_ : Union[str, Any] = model.reverse_bettertransformer() self.assertFalse(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__magic_name__ ) snake_case_ : str = AutoModelForSeqaSeqLM.from_pretrained(__magic_name__ ) self.assertFalse( any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) snake_case_ : Tuple = model_reloaded.generate(**__magic_name__ ) self.assertTrue(torch.allclose(__magic_name__ , __magic_name__ ) ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Optional[int] = '''hf-internal-testing/tiny-random-t5''' snake_case_ : str = AutoModelForSeqaSeqLM.from_pretrained(__magic_name__ ) snake_case_ : str = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(__magic_name__ ): model.save_pretrained(__magic_name__ ) snake_case_ : List[Any] = model.reverse_bettertransformer() model.save_pretrained(__magic_name__ )
279
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" _UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ), F"""{len(__lowerCAmelCase )} != {len(__lowerCAmelCase )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) _a = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _a = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def __A ( __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" try: _UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(__lowerCAmelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __A ( __lowerCAmelCase , __lowerCAmelCase = "student" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , )-> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" _UpperCAmelCase = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowerCAmelCase , __lowerCAmelCase ): AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) # purely for convenience _UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(__lowerCAmelCase ).eval() else: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), F"""teacher must be a model or string got type {type(__lowerCAmelCase )}""" _UpperCAmelCase = teacher.config.to_diff_dict() try: _UpperCAmelCase , _UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowerCAmelCase ) # Copy weights _UpperCAmelCase = teacher.config_class(**__lowerCAmelCase ) _UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. _UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=__lowerCAmelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save _UpperCAmelCase , _UpperCAmelCase = list(range(__lowerCAmelCase ) ), list(range(__lowerCAmelCase ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(__lowerCAmelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) if d_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) try: if hasattr( __lowerCAmelCase , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowerCAmelCase ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowerCAmelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowerCAmelCase ) copy_layers(teacher.decoder.block , student.decoder.block , __lowerCAmelCase ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) _UpperCAmelCase = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowerCAmelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
39
0
from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP lowercase__ =logging.get_logger(__name__) # pylint: disable=invalid-name lowercase__ ='\n Examples:\n ```py\n >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyPriorPipeline.from_pretrained("kandinsky-community/Kandinsky-2-1-prior")\n >>> pipe_prior.to("cuda")\n\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> negative_image_emb = out.negative_image_embeds\n\n >>> pipe = KandinskyPipeline.from_pretrained("kandinsky-community/kandinsky-2-1")\n >>> pipe.to("cuda")\n\n >>> image = pipe(\n ... prompt,\n ... image_embeds=image_emb,\n ... negative_image_embeds=negative_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... ).images\n\n >>> image[0].save("cat.png")\n ```\n' def __UpperCamelCase ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Optional[Any]=8 ): __a : str = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 __a : List[str] = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class UpperCamelCase__ ( snake_case__ ): def __init__(self : Union[str, Any] , snake_case_ : Any , snake_case_ : int , snake_case_ : Optional[Any] , snake_case_ : Any , snake_case_ : Union[str, Any] , ): super().__init__() self.register_modules( text_encoder=snake_case_ , tokenizer=snake_case_ , unet=snake_case_ , scheduler=snake_case_ , movq=snake_case_ , ) __a : Optional[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase (self : Optional[Any] , snake_case_ : Tuple , snake_case_ : int , snake_case_ : int , snake_case_ : Tuple , snake_case_ : List[str] , snake_case_ : str ): if latents is None: __a : Optional[int] = randn_tensor(snake_case_ , generator=snake_case_ , device=snake_case_ , dtype=snake_case_ ) else: if latents.shape != shape: raise ValueError(f"Unexpected latents shape, got {latents.shape}, expected {shape}" ) __a : Union[str, Any] = latents.to(snake_case_ ) __a : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase (self : str , snake_case_ : List[str] , snake_case_ : str , snake_case_ : Dict , snake_case_ : str , snake_case_ : Optional[int]=None , ): __a : Any = len(snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else 1 # get prompt text embeddings __a : Tuple = self.tokenizer( snake_case_ , padding='''max_length''' , truncation=snake_case_ , max_length=7_7 , return_attention_mask=snake_case_ , add_special_tokens=snake_case_ , return_tensors='''pt''' , ) __a : Tuple = text_inputs.input_ids __a : List[str] = self.tokenizer(snake_case_ , padding='''longest''' , return_tensors='''pt''' ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(snake_case_ , snake_case_ ): __a : Dict = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' f" {self.tokenizer.model_max_length} tokens: {removed_text}" ) __a : Tuple = text_input_ids.to(snake_case_ ) __a : str = text_inputs.attention_mask.to(snake_case_ ) __a , __a : Tuple = self.text_encoder( input_ids=snake_case_ , attention_mask=snake_case_ ) __a : Optional[Any] = prompt_embeds.repeat_interleave(snake_case_ , dim=0 ) __a : str = text_encoder_hidden_states.repeat_interleave(snake_case_ , dim=0 ) __a : Dict = text_mask.repeat_interleave(snake_case_ , dim=0 ) if do_classifier_free_guidance: __a : Tuple = 4_2 if negative_prompt is None: __a : int = [''''''] * batch_size elif type(snake_case_ ) is not type(snake_case_ ): raise TypeError( f"`negative_prompt` should be the same type to `prompt`, but got {type(snake_case_ )} !=" f" {type(snake_case_ )}." ) elif isinstance(snake_case_ , snake_case_ ): __a : str = [negative_prompt] elif batch_size != len(snake_case_ ): raise ValueError( f"`negative_prompt`: {negative_prompt} has batch size {len(snake_case_ )}, but `prompt`:" f" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches" ''' the batch size of `prompt`.''' ) else: __a : Dict = negative_prompt __a : Dict = self.tokenizer( snake_case_ , padding='''max_length''' , max_length=7_7 , truncation=snake_case_ , return_attention_mask=snake_case_ , add_special_tokens=snake_case_ , return_tensors='''pt''' , ) __a : Optional[Any] = uncond_input.input_ids.to(snake_case_ ) __a : Dict = uncond_input.attention_mask.to(snake_case_ ) __a , __a : int = self.text_encoder( input_ids=snake_case_ , attention_mask=snake_case_ ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __a : List[Any] = negative_prompt_embeds.shape[1] __a : str = negative_prompt_embeds.repeat(1 , snake_case_ ) __a : Union[str, Any] = negative_prompt_embeds.view(batch_size * num_images_per_prompt , snake_case_ ) __a : str = uncond_text_encoder_hidden_states.shape[1] __a : Any = uncond_text_encoder_hidden_states.repeat(1 , snake_case_ , 1 ) __a : List[str] = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , snake_case_ , -1 ) __a : List[str] = uncond_text_mask.repeat_interleave(snake_case_ , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __a : Optional[int] = torch.cat([negative_prompt_embeds, prompt_embeds] ) __a : List[Any] = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) __a : List[str] = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def lowerCAmelCase (self : List[str] , snake_case_ : int=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) __a : Dict = torch.device(f"cuda:{gpu_id}" ) __a : Optional[Any] = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(snake_case_ , snake_case_ ) def lowerCAmelCase (self : Optional[Any] , snake_case_ : Dict=0 ): if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) __a : List[Any] = torch.device(f"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=snake_case_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __a : List[Any] = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: __a , __a : List[str] = cpu_offload_with_hook(snake_case_ , snake_case_ , prev_module_hook=snake_case_ ) if self.safety_checker is not None: __a , __a : Tuple = cpu_offload_with_hook(self.safety_checker , snake_case_ , prev_module_hook=snake_case_ ) # We'll offload the last model manually. __a : Optional[Any] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase (self : List[Any] ): if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(snake_case_ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(snake_case_ ) def __call__(self : Optional[Any] , snake_case_ : str , snake_case_ : int , snake_case_ : str , snake_case_ : Any = None , snake_case_ : List[Any] = 5_1_2 , snake_case_ : Union[str, Any] = 5_1_2 , snake_case_ : str = 1_0_0 , snake_case_ : Union[str, Any] = 4.0 , snake_case_ : Tuple = 1 , snake_case_ : Dict = None , snake_case_ : Optional[Any] = None , snake_case_ : Optional[int] = "pil" , snake_case_ : List[str] = True , ): if isinstance(snake_case_ , snake_case_ ): __a : Any = 1 elif isinstance(snake_case_ , snake_case_ ): __a : List[Any] = len(snake_case_ ) else: raise ValueError(f"`prompt` has to be of type `str` or `list` but is {type(snake_case_ )}" ) __a : int = self._execution_device __a : Union[str, Any] = batch_size * num_images_per_prompt __a : Dict = guidance_scale > 1.0 __a , __a , __a : int = self._encode_prompt( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if isinstance(snake_case_ , snake_case_ ): __a : Optional[Any] = torch.cat(snake_case_ , dim=0 ) if isinstance(snake_case_ , snake_case_ ): __a : List[str] = torch.cat(snake_case_ , dim=0 ) if do_classifier_free_guidance: __a : int = image_embeds.repeat_interleave(snake_case_ , dim=0 ) __a : List[Any] = negative_image_embeds.repeat_interleave(snake_case_ , dim=0 ) __a : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=snake_case_ ) self.scheduler.set_timesteps(snake_case_ , device=snake_case_ ) __a : Union[str, Any] = self.scheduler.timesteps __a : Tuple = self.unet.config.in_channels __a , __a : Any = get_new_h_w(snake_case_ , snake_case_ , self.movq_scale_factor ) # create initial latent __a : Union[str, Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , snake_case_ , snake_case_ , snake_case_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(snake_case_ ) ): # expand the latents if we are doing classifier free guidance __a : List[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __a : int = {'''text_embeds''': prompt_embeds, '''image_embeds''': image_embeds} __a : List[Any] = self.unet( sample=snake_case_ , timestep=snake_case_ , encoder_hidden_states=snake_case_ , added_cond_kwargs=snake_case_ , return_dict=snake_case_ , )[0] if do_classifier_free_guidance: __a , __a : List[str] = noise_pred.split(latents.shape[1] , dim=1 ) __a , __a : List[Any] = noise_pred.chunk(2 ) __a , __a : List[Any] = variance_pred.chunk(2 ) __a : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __a : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __a , __a : str = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __a : Dict = self.scheduler.step( snake_case_ , snake_case_ , snake_case_ , generator=snake_case_ , ).prev_sample # post-processing __a : List[Any] = self.movq.decode(snake_case_ , force_not_quantize=snake_case_ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: __a : Optional[int] = image * 0.5 + 0.5 __a : List[Any] = image.clamp(0 , 1 ) __a : int = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __a : Optional[int] = self.numpy_to_pil(snake_case_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=snake_case_ )
216
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False )-> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase = '' else: _UpperCAmelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase = in_proj_bias[: config.hidden_size] _UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase = in_proj_bias[-config.hidden_size :] def __A ( __lowerCAmelCase )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" _UpperCAmelCase = dct.pop(__lowerCAmelCase ) _UpperCAmelCase = val def __A ( )-> str: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True )-> List[str]: """simple docstring""" _UpperCAmelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCAmelCase = 8 # set labels if required if not base_model: _UpperCAmelCase = 1_000 _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCAmelCase = 384 _UpperCAmelCase = 1_536 _UpperCAmelCase = 12 _UpperCAmelCase = 6 # load original model from torch hub _UpperCAmelCase = torch.hub.load('facebookresearch/dino:main' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCAmelCase = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _UpperCAmelCase = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _UpperCAmelCase = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCAmelCase = ViTImageProcessor() _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCAmelCase = encoding['pixel_values'] _UpperCAmelCase = model(__lowerCAmelCase ) if base_model: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) _a = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
39
0
import numpy as np _lowerCAmelCase : Optional[Any] = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class __magic_name__ : def __init__( self ) -> Optional[Any]: '''simple docstring''' __a =np.array(__snake_case ) def __magic_name__ ( self , __snake_case ) -> Any: '''simple docstring''' __a , __a =np.where(letter == self.SQUARE ) __a =np.concatenate([indexa + 1, indexa + 1] ) return indexes def __magic_name__ ( self , __snake_case , __snake_case ) -> List[Any]: '''simple docstring''' __a =self.SQUARE[indexa - 1, indexa - 1] return letter def __magic_name__ ( self , __snake_case ) -> Optional[int]: '''simple docstring''' __a =message.lower() __a =message.replace(' ' , '' ) __a =message.replace('j' , 'i' ) __a =np.empty((2, len(__snake_case )) ) for letter_index in range(len(__snake_case ) ): __a =self.letter_to_numbers(message[letter_index] ) __a =numbers[0] __a =numbers[1] __a =first_step.reshape(2 * len(__snake_case ) ) __a ='' for numbers_index in range(len(__snake_case ) ): __a =int(second_step[numbers_index * 2] ) __a =int(second_step[(numbers_index * 2) + 1] ) __a =self.numbers_to_letter(__snake_case , __snake_case ) __a =encoded_message + letter return encoded_message def __magic_name__ ( self , __snake_case ) -> Tuple: '''simple docstring''' __a =message.lower() message.replace(' ' , '' ) __a =np.empty(2 * len(__snake_case ) ) for letter_index in range(len(__snake_case ) ): __a =self.letter_to_numbers(message[letter_index] ) __a =numbers[0] __a =numbers[1] __a =first_step.reshape((2, len(__snake_case )) ) __a ='' for numbers_index in range(len(__snake_case ) ): __a =int(second_step[0, numbers_index] ) __a =int(second_step[1, numbers_index] ) __a =self.numbers_to_letter(__snake_case , __snake_case ) __a =decoded_message + letter return decoded_message
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
from math import pi, sqrt, tan def _UpperCamelCase ( snake_case__ ) -> float: if side_length < 0: raise ValueError("surface_area_cube() only accepts non-negative values" ) return 6 * side_length**2 def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> float: if length < 0 or breadth < 0 or height < 0: raise ValueError("surface_area_cuboid() only accepts non-negative values" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def _UpperCamelCase ( snake_case__ ) -> float: if radius < 0: raise ValueError("surface_area_sphere() only accepts non-negative values" ) return 4 * pi * radius**2 def _UpperCamelCase ( snake_case__ ) -> float: if radius < 0: raise ValueError("surface_area_hemisphere() only accepts non-negative values" ) return 3 * pi * radius**2 def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if radius < 0 or height < 0: raise ValueError("surface_area_cone() only accepts non-negative values" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( "surface_area_conical_frustum() only accepts non-negative values" ) __UpperCAmelCase : Union[str, Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if radius < 0 or height < 0: raise ValueError("surface_area_cylinder() only accepts non-negative values" ) return 2 * pi * radius * (height + radius) def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if torus_radius < 0 or tube_radius < 0: raise ValueError("surface_area_torus() only accepts non-negative values" ) if torus_radius < tube_radius: raise ValueError( "surface_area_torus() does not support spindle or self intersecting tori" ) return 4 * pow(__lowerCAmelCase, 2 ) * torus_radius * tube_radius def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if length < 0 or width < 0: raise ValueError("area_rectangle() only accepts non-negative values" ) return length * width def _UpperCamelCase ( snake_case__ ) -> float: if side_length < 0: raise ValueError("area_square() only accepts non-negative values" ) return side_length**2 def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if base < 0 or height < 0: raise ValueError("area_triangle() only accepts non-negative values" ) return (base * height) / 2 def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("area_triangle_three_sides() only accepts non-negative values" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("Given three sides do not form a triangle" ) __UpperCAmelCase : Dict = (sidea + sidea + sidea) / 2 __UpperCAmelCase : Optional[Any] = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if base < 0 or height < 0: raise ValueError("area_parallelogram() only accepts non-negative values" ) return base * height def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> float: if basea < 0 or basea < 0 or height < 0: raise ValueError("area_trapezium() only accepts non-negative values" ) return 1 / 2 * (basea + basea) * height def _UpperCamelCase ( snake_case__ ) -> float: if radius < 0: raise ValueError("area_circle() only accepts non-negative values" ) return pi * radius**2 def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if radius_x < 0 or radius_y < 0: raise ValueError("area_ellipse() only accepts non-negative values" ) return pi * radius_x * radius_y def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError("area_rhombus() only accepts non-negative values" ) return 1 / 2 * diagonal_a * diagonal_a def _UpperCamelCase ( snake_case__, snake_case__ ) -> float: if not isinstance(__lowerCAmelCase, __lowerCAmelCase ) or sides < 3: raise ValueError( "area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides" ) elif length < 0: raise ValueError( "area_reg_polygon() only accepts non-negative values as \ length of a side" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F'Rectangle: {area_rectangle(10, 20) = }') print(F'Square: {area_square(10) = }') print(F'Triangle: {area_triangle(10, 10) = }') print(F'Triangle: {area_triangle_three_sides(5, 12, 13) = }') print(F'Parallelogram: {area_parallelogram(10, 20) = }') print(F'Rhombus: {area_rhombus(10, 20) = }') print(F'Trapezium: {area_trapezium(10, 20, 30) = }') print(F'Circle: {area_circle(20) = }') print(F'Ellipse: {area_ellipse(10, 20) = }') print('''\nSurface Areas of various geometric shapes: \n''') print(F'Cube: {surface_area_cube(20) = }') print(F'Cuboid: {surface_area_cuboid(10, 20, 30) = }') print(F'Sphere: {surface_area_sphere(20) = }') print(F'Hemisphere: {surface_area_hemisphere(20) = }') print(F'Cone: {surface_area_cone(10, 20) = }') print(F'Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }') print(F'Cylinder: {surface_area_cylinder(10, 20) = }') print(F'Torus: {surface_area_torus(20, 10) = }') print(F'Equilateral Triangle: {area_reg_polygon(3, 10) = }') print(F'Square: {area_reg_polygon(4, 10) = }') print(F'Reqular Pentagon: {area_reg_polygon(5, 10) = }')
157
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=3 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=10 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[1, 1, 2, 1] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase="relu" , UpperCAmelCase=3 , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = embeddings_size _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = scope _UpperCAmelCase = len(UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = TFResNetModel(config=UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFResNetForImageClassification(UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase = layer_type _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) @slow def UpperCamelCase ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TFResNetModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __A ( )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" @cached_property def UpperCamelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=UpperCAmelCase , return_tensors='tf' ) # forward pass _UpperCAmelCase = model(**UpperCAmelCase ) # verify the logits _UpperCAmelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _UpperCAmelCase = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCAmelCase , atol=1e-4 ) )
39
0
'''simple docstring''' import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=99 , _lowerCAmelCase=13 , _lowerCAmelCase=16 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=2 , _lowerCAmelCase=32 , _lowerCAmelCase=4 , _lowerCAmelCase=4 , _lowerCAmelCase=30 , _lowerCAmelCase=0 , _lowerCAmelCase=1 , _lowerCAmelCase=2 , _lowerCAmelCase=None , ) -> List[Any]: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = decoder_seq_length # For common tests _lowerCAmelCase = self.decoder_seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_attention_mask _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = d_model _lowerCAmelCase = d_model _lowerCAmelCase = decoder_layers _lowerCAmelCase = decoder_layers _lowerCAmelCase = decoder_ffn_dim _lowerCAmelCase = decoder_attention_heads _lowerCAmelCase = decoder_attention_heads _lowerCAmelCase = eos_token_id _lowerCAmelCase = bos_token_id _lowerCAmelCase = pad_token_id _lowerCAmelCase = decoder_start_token_id _lowerCAmelCase = use_cache _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = None _lowerCAmelCase = decoder_seq_length _lowerCAmelCase = 2 _lowerCAmelCase = 1 def _snake_case ( self ) -> Tuple: _lowerCAmelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCAmelCase = None if self.use_attention_mask: _lowerCAmelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCAmelCase = None if self.use_labels: _lowerCAmelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCAmelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> Optional[int]: _lowerCAmelCase = True _lowerCAmelCase = TrOCRDecoder(config=_lowerCAmelCase ).to(_lowerCAmelCase ).eval() _lowerCAmelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCAmelCase = model(_lowerCAmelCase , use_cache=_lowerCAmelCase ) _lowerCAmelCase = model(_lowerCAmelCase ) _lowerCAmelCase = model(_lowerCAmelCase , use_cache=_lowerCAmelCase ) self.parent.assertTrue(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) ) self.parent.assertTrue(len(_lowerCAmelCase ) == len(_lowerCAmelCase ) + 1 ) _lowerCAmelCase = outputs["past_key_values"] # create hypothetical next token and extent to next_input_ids _lowerCAmelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCAmelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCAmelCase = model(_lowerCAmelCase )["last_hidden_state"] _lowerCAmelCase = model(_lowerCAmelCase , past_key_values=_lowerCAmelCase )["last_hidden_state"] # select random slice _lowerCAmelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCAmelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCAmelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-3 ) def _snake_case ( self ) -> Optional[Any]: _lowerCAmelCase = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = config_and_inputs _lowerCAmelCase = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict @require_torch class lowerCAmelCase_ ( snake_case__ ,snake_case__ ,snake_case__ ,unittest.TestCase ): __lowerCamelCase : Dict = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () __lowerCamelCase : Optional[int] = (TrOCRForCausalLM,) if is_torch_available() else () __lowerCamelCase : Any = {"text-generation": TrOCRForCausalLM} if is_torch_available() else {} __lowerCamelCase : List[Any] = True __lowerCamelCase : str = False def _snake_case ( self ) -> Dict: _lowerCAmelCase = TrOCRStandaloneDecoderModelTester(self , is_training=_lowerCAmelCase ) _lowerCAmelCase = ConfigTester(self , config_class=_lowerCAmelCase ) def _snake_case ( self ) -> List[str]: pass def _snake_case ( self ) -> List[str]: pass def _snake_case ( self ) -> List[Any]: pass def _snake_case ( self ) -> Dict: self.config_tester.run_common_tests() def _snake_case ( self ) -> List[Any]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*_lowerCAmelCase ) def _snake_case ( self ) -> Any: return @unittest.skip("The model doesn\'t support left padding" ) # and it's not used enough to be worth fixing :) def _snake_case ( self ) -> Tuple: pass
158
def __A ( __lowerCAmelCase )-> list: """simple docstring""" if len(__lowerCAmelCase ) < 2: return collection def circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , mid + 1 , __lowerCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) - 1 ) return collection if __name__ == "__main__": _a = input('''Enter numbers separated by a comma:\n''').strip() _a = [int(item) for item in user_input.split(''',''')] print(circle_sort(unsorted))
39
0
from __future__ import annotations lowerCamelCase__ = { """A""": ["""B""", """C""", """E"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F""", """G"""], """D""": ["""B"""], """E""": ["""A""", """B""", """D"""], """F""": ["""C"""], """G""": ["""C"""], } class A__ : def __init__( self : List[str] , a : Dict , a : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = graph # mapping node to its parent in resulting breadth first tree lowerCAmelCase__ : List[str] = {} lowerCAmelCase__ : int = source_vertex def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Any = {self.source_vertex} lowerCAmelCase__ : int = None lowerCAmelCase__ : Dict = [self.source_vertex] # first in first out queue while queue: lowerCAmelCase__ : int = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(a ) lowerCAmelCase__ : Any = vertex queue.append(a ) def _lowerCamelCase ( self : str , a : str ): '''simple docstring''' if target_vertex == self.source_vertex: return self.source_vertex lowerCAmelCase__ : Optional[int] = self.parent.get(a ) if target_vertex_parent is None: lowerCAmelCase__ : Any = ( f'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(a ) return self.shortest_path(a ) + f'''->{target_vertex}''' if __name__ == "__main__": lowerCamelCase__ = Graph(graph, """G""") g.breath_first_search() print(g.shortest_path("""D""")) print(g.shortest_path("""G""")) print(g.shortest_path("""Foo"""))
212
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "Pix2StructImageProcessor" UpperCamelCase__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = False super().__init__(UpperCAmelCase , UpperCAmelCase ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 2048 , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , **UpperCAmelCase ) else: # add pixel_values and bbox _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , header_text=UpperCAmelCase , **UpperCAmelCase ) if text is not None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) if "attention_mask" in text_encoding: _UpperCAmelCase = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: _UpperCAmelCase = text_encoding.pop('input_ids' ) else: _UpperCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase ) return encoding_image_processor def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCAmelCase ( snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' @register_to_config def __init__( self : Optional[int] ,A : str ,A : Dict ,A : Any ,A : Any ,A : Optional[int] ,A : List[str] ,A : Tuple ,A : int ,A : int ,A : Dict = False ,): super().__init__() __A = nn.Embedding(A ,A ) __A = nn.Embedding(A ,A ) __A = False __A = nn.Dropout(p=A ) __A = TaConfig( vocab_size=A ,d_model=A ,num_heads=A ,d_kv=A ,d_ff=A ,dropout_rate=A ,feed_forward_proj=A ,is_decoder=A ,is_encoder_decoder=A ,) __A = nn.ModuleList() for lyr_num in range(A ): __A = TaBlock(A ) self.encoders.append(A ) __A = TaLayerNorm(A ) __A = nn.Dropout(p=A ) def UpperCamelCase_ ( self : Optional[Any] ,A : Optional[int] ,A : Dict ): __A = self.token_embedder(A ) __A = encoder_input_tokens.shape[1] __A = torch.arange(A ,device=encoder_input_tokens.device ) x += self.position_encoding(A ) __A = self.dropout_pre(A ) # inverted the attention mask __A = encoder_input_tokens.size() __A = self.get_extended_attention_mask(A ,A ) for lyr in self.encoders: __A = lyr(A ,A )[0] __A = self.layer_norm(A ) return self.dropout_post(A ), encoder_inputs_mask
15
class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase = "" , UpperCAmelCase = False ): """simple docstring""" _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 0 for q, w in zip(self.prefix , UpperCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=UpperCAmelCase , is_leaf=UpperCAmelCase ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def UpperCamelCase ( self , UpperCAmelCase = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = RadixNode() root.insert_many(__lowerCAmelCase ) assert all(root.find(__lowerCAmelCase ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = 'banana bananas bandanas bandana band apple all beast'.split() root.insert_many(__lowerCAmelCase ) print('Words:' , __lowerCAmelCase ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
39
0
def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Tuple = 60_08_51_47_51_43 ) -> int: """simple docstring""" try: SCREAMING_SNAKE_CASE__ = int(__lowerCAmelCase ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 SCREAMING_SNAKE_CASE__ = i while n % i == 0: SCREAMING_SNAKE_CASE__ = n // i i += 1 return int(__lowerCAmelCase ) if __name__ == "__main__": print(F"""{solution() = }""")
219
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class __lowerCamelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments UpperCAmelCase="<s>" , UpperCAmelCase="<pad>" , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = bos, unk, pad, eos _UpperCAmelCase = [] _UpperCAmelCase = [] _UpperCAmelCase = {} _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , UpperCAmelCase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , UpperCAmelCase ): """simple docstring""" return sym in self.indices @classmethod def UpperCamelCase ( cls , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = cls() d.add_from_file(UpperCAmelCase ) return d def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase=1 , UpperCAmelCase=False ): """simple docstring""" if word in self.indices and not overwrite: _UpperCAmelCase = self.indices[word] _UpperCAmelCase = self.count[idx] + n return idx else: _UpperCAmelCase = len(self.symbols ) _UpperCAmelCase = idx self.symbols.append(UpperCAmelCase ) self.count.append(UpperCAmelCase ) return idx def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return 0 def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(UpperCAmelCase ) ) return _UpperCAmelCase = f.readlines() _UpperCAmelCase = self._load_meta(UpperCAmelCase ) for line in lines[indices_start_line:]: try: _UpperCAmelCase , _UpperCAmelCase = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": _UpperCAmelCase = True _UpperCAmelCase , _UpperCAmelCase = line.rsplit(' ' , 1 ) else: _UpperCAmelCase = False _UpperCAmelCase = int(UpperCAmelCase ) _UpperCAmelCase = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(UpperCAmelCase ) ) self.add_symbol(UpperCAmelCase , n=UpperCAmelCase , overwrite=UpperCAmelCase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" _UpperCAmelCase = dict((re.sub(R'@@$' , '' , __lowerCAmelCase ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , __lowerCAmelCase ), v) for k, v in d.items() ) _UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] _UpperCAmelCase = d[k] # restore return da def __A ( __lowerCAmelCase , __lowerCAmelCase )-> str: """simple docstring""" if not os.path.exists(__lowerCAmelCase ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'checkpoint.pt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location='cpu' ) _UpperCAmelCase = chkpt['cfg']['model'] # dicts _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'dict.txt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) _UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) _UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) _UpperCAmelCase = len(__lowerCAmelCase ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['vocab_file'] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # merges_file (bpecodes) _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'bpecodes' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowerCAmelCase , __lowerCAmelCase ) # model config _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'config.json' ) _UpperCAmelCase = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1E-12, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # tokenizer config _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 1_024, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # model _UpperCAmelCase = chkpt['model'] # remove unneeded keys _UpperCAmelCase = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) else: _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase = BioGptConfig.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase = BioGptForCausalLM(__lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(__lowerCAmelCase ) # save _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) print('Conversion is done!' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
39
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore lowerCAmelCase__ = """ Human: <<task>> Assistant: """ lowerCAmelCase__ = """huggingface-tools/default-prompts""" lowerCAmelCase__ = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Optional[Any] , SCREAMING_SNAKE_CASE_: Tuple , SCREAMING_SNAKE_CASE_: str="run" ) -> Dict: '''simple docstring''' if prompt_or_repo_id is None: A__ = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , __lowerCAmelCase ) is not None: return prompt_or_repo_id A__ = cached_file( __lowerCAmelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(__lowerCAmelCase , "r" , encoding="utf-8" ) as f: return f.read()
68
from __future__ import annotations import collections import pprint from pathlib import Path def __A ( __lowerCAmelCase )-> str: """simple docstring""" return "".join(sorted(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase )-> list[str]: """simple docstring""" return word_by_signature[signature(__lowerCAmelCase )] _a = Path(__file__).parent.joinpath('''words.txt''').read_text(encoding='''utf-8''') _a = sorted({word.strip().lower() for word in data.splitlines()}) _a = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _a = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('''anagrams.txt''', '''w''') as file: file.write('''all_anagrams = \n ''') file.write(pprint.pformat(all_anagrams))
39
0
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase__ : Optional[int] = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Union[str, Any] = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys UpperCAmelCase__ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
121
from __future__ import annotations def __A ( __lowerCAmelCase )-> list[int]: """simple docstring""" _UpperCAmelCase = 2 _UpperCAmelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCAmelCase ) if n > 1: factors.append(__lowerCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
39
0
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case, __snake_case = None, __snake_case = None, __snake_case = False, ) -> tuple[int, float, str]: """simple docstring""" _UpperCamelCase = cipher_alphabet or [chr(__lowerCAmelCase ) for i in range(97, 1_23 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) _UpperCamelCase = { '''a''': 0.08497, '''b''': 0.01492, '''c''': 0.02202, '''d''': 0.04253, '''e''': 0.11162, '''f''': 0.02228, '''g''': 0.02015, '''h''': 0.06094, '''i''': 0.07546, '''j''': 0.00153, '''k''': 0.01292, '''l''': 0.04025, '''m''': 0.02406, '''n''': 0.06749, '''o''': 0.07507, '''p''': 0.01929, '''q''': 0.00095, '''r''': 0.07587, '''s''': 0.06327, '''t''': 0.09356, '''u''': 0.02758, '''v''': 0.00978, '''w''': 0.02560, '''x''': 0.00150, '''y''': 0.01994, '''z''': 0.00077, } else: # Custom frequencies dictionary _UpperCamelCase = frequencies_dict if not case_sensitive: _UpperCamelCase = ciphertext.lower() # Chi squared statistic values _UpperCamelCase = {} # cycle through all of the shifts for shift in range(len(__lowerCAmelCase ) ): _UpperCamelCase = '''''' # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet _UpperCamelCase = (alphabet_letters.index(letter.lower() ) - shift) % len( __lowerCAmelCase ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter _UpperCamelCase = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: _UpperCamelCase = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message _UpperCamelCase = decrypted_with_shift.lower().count(__lowerCAmelCase ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _UpperCamelCase = frequencies[letter] * occurrences # Complete the chi squared statistic formula _UpperCamelCase = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message _UpperCamelCase = decrypted_with_shift.count(__lowerCAmelCase ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _UpperCamelCase = frequencies[letter] * occurrences # Complete the chi squared statistic formula _UpperCamelCase = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary _UpperCamelCase = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(__snake_case ) -> tuple[float, str]: return chi_squared_statistic_values[key] _UpperCamelCase = min( __lowerCAmelCase, key=__lowerCAmelCase, ) # Get all the data from the most likely cipher (key, decoded message) ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
194
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __A ( )-> tuple[list[int], int]: """simple docstring""" _UpperCAmelCase = [randint(-1_000 , 1_000 ) for i in range(10 )] _UpperCAmelCase = randint(-5_000 , 5_000 ) return (arr, r) _a = make_dataset() def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, ...]: """simple docstring""" for triplet in permutations(__lowerCAmelCase , 3 ): if sum(__lowerCAmelCase ) == target: return tuple(sorted(__lowerCAmelCase ) ) return (0, 0, 0) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> tuple[int, int, int]: """simple docstring""" arr.sort() _UpperCAmelCase = len(__lowerCAmelCase ) for i in range(n - 1 ): _UpperCAmelCase , _UpperCAmelCase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __A ( )-> tuple[float, float]: """simple docstring""" _UpperCAmelCase = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' _UpperCAmelCase = '\ntriplet_sum1(*dataset)\n' _UpperCAmelCase = '\ntriplet_sum2(*dataset)\n' _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) _UpperCAmelCase = repeat(setup=__lowerCAmelCase , stmt=__lowerCAmelCase , repeat=5 , number=10_000 ) return (min(__lowerCAmelCase ), min(__lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _a = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
39
0
import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class __lowerCAmelCase ( snake_case__ ): def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def lowerCamelCase (self ) -> str: '''simple docstring''' with self.assertRaises(__magic_name__ ): snake_case_ : Optional[int] = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' with self.assertRaises(__magic_name__ ): snake_case_ : List[Any] = pa.array(TypedSequence([1, 2, 3] , try_type=Value('''bool''' ) , type=Value('''int64''' ) ) ) def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[int] = pa.array(TypedSequence([1, 2, 3] , type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): snake_case_ : str = pa.array(TypedSequence(['''foo''', '''bar'''] , type=Value('''int64''' ) ) ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : List[str] = pa.array(TypedSequence([1, 2, 3] , try_type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=Value('''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def lowerCamelCase (self ) -> Tuple: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): snake_case_ : Optional[int] = pa.array(TypedSequence(['''foo''', '''bar'''] , type=ArrayaD((1, 3) , '''int64''' ) ) ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : Tuple = pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' import PIL.Image snake_case_ : Dict = PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( '''datasets.arrow_writer.cast_to_python_objects''' , side_effect=__magic_name__ ) as mock_cast_to_python_objects: snake_case_ : str = pa.array(TypedSequence([{'''path''': None, '''bytes''': B'''image_bytes'''}, pil_image] , type=Image() ) ) snake_case_ , snake_case_ : List[str] = mock_cast_to_python_objects.call_args_list[-1] self.assertIn('''optimize_list_casting''' , __magic_name__ ) self.assertFalse(kwargs['''optimize_list_casting'''] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> List[str]: """simple docstring""" snake_case_ : Union[str, Any] = pa.BufferReader(__lowerCAmelCase ) if isinstance(__lowerCAmelCase , pa.Buffer ) else pa.memory_map(__lowerCAmelCase ) snake_case_ : Optional[int] = pa.ipc.open_stream(__lowerCAmelCase ) snake_case_ : Optional[int] = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 10] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> List[str]: """simple docstring""" snake_case_ : int = pa.BufferOutputStream() snake_case_ : Union[str, Any] = pa.schema(__lowerCAmelCase ) if fields else None with ArrowWriter(stream=__lowerCAmelCase , schema=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) snake_case_ , snake_case_ : Optional[int] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: snake_case_ : int = {'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__lowerCAmelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def lowerCamelCase_ ( ) -> Union[str, Any]: """simple docstring""" snake_case_ : Optional[Any] = pa.BufferOutputStream() snake_case_ : Optional[int] = Features({'''labels''': ClassLabel(names=['''neg''', '''pos'''] )} ) with ArrowWriter(stream=__lowerCAmelCase , features=__lowerCAmelCase ) as writer: writer.write({'''labels''': 0} ) writer.write({'''labels''': 1} ) snake_case_ , snake_case_ : List[Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata snake_case_ : Optional[Any] = pa.BufferReader(output.getvalue() ) snake_case_ : Tuple = pa.ipc.open_stream(__lowerCAmelCase ) snake_case_ : List[Any] = f.read_all() snake_case_ : Tuple = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(__lowerCAmelCase ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 10] ) def lowerCamelCase_ ( _UpperCamelCase ) -> Dict: """simple docstring""" snake_case_ : Optional[int] = pa.BufferOutputStream() with ArrowWriter( stream=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase , hash_salt='''split_name''' , check_duplicates=__lowerCAmelCase , ) as writer: with pytest.raises(__lowerCAmelCase ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=[1, 2] ) snake_case_ , snake_case_ : Any = writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 10] ) def lowerCamelCase_ ( _UpperCamelCase ) -> Any: """simple docstring""" snake_case_ : Tuple = pa.BufferOutputStream() with ArrowWriter( stream=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase , hash_salt='''split_name''' , check_duplicates=__lowerCAmelCase , ) as writer: with pytest.raises(__lowerCAmelCase ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=10 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=10 ) snake_case_ , snake_case_ : Union[str, Any] = writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 10] ) def lowerCamelCase_ ( _UpperCamelCase ) -> Any: """simple docstring""" snake_case_ : Optional[int] = pa.BufferOutputStream() with ArrowWriter( stream=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase , hash_salt='''split_name''' , check_duplicates=__lowerCAmelCase , ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=1 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=2 ) snake_case_ , snake_case_ : Union[str, Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 10] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Optional[int]: """simple docstring""" snake_case_ : Dict = pa.BufferOutputStream() snake_case_ : Union[str, Any] = pa.schema(__lowerCAmelCase ) if fields else None with ArrowWriter(stream=__lowerCAmelCase , schema=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) writer.write_batch({'''col_1''': [], '''col_2''': []} ) snake_case_ , snake_case_ : Tuple = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: snake_case_ : List[str] = {'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__lowerCAmelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 10] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" snake_case_ : int = pa.BufferOutputStream() snake_case_ : Optional[int] = pa.schema(__lowerCAmelCase ) if fields else None with ArrowWriter(stream=__lowerCAmelCase , schema=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase ) as writer: writer.write_table(pa.Table.from_pydict({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) ) snake_case_ , snake_case_ : Tuple = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: snake_case_ : Optional[Any] = {'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__lowerCAmelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 10] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> List[str]: """simple docstring""" snake_case_ : int = pa.BufferOutputStream() snake_case_ : List[str] = pa.schema(__lowerCAmelCase ) if fields else None with ArrowWriter(stream=__lowerCAmelCase , schema=__lowerCAmelCase , writer_batch_size=__lowerCAmelCase ) as writer: writer.write_row(pa.Table.from_pydict({'''col_1''': ['''foo'''], '''col_2''': [1]} ) ) writer.write_row(pa.Table.from_pydict({'''col_1''': ['''bar'''], '''col_2''': [2]} ) ) snake_case_ , snake_case_ : Tuple = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: snake_case_ : Union[str, Any] = {'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__lowerCAmelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def lowerCamelCase_ ( ) -> int: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ : int = {'''col_1''': pa.string(), '''col_2''': pa.intaa()} snake_case_ : List[str] = os.path.join(__lowerCAmelCase , '''test.arrow''' ) with ArrowWriter(path=__lowerCAmelCase , schema=pa.schema(__lowerCAmelCase ) ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) snake_case_ , snake_case_ : Optional[Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(__lowerCAmelCase , metadata=writer._schema.metadata ) _check_output(__lowerCAmelCase , 1 ) def lowerCamelCase_ ( _UpperCamelCase ) -> Tuple: """simple docstring""" if pa.types.is_list(__lowerCAmelCase ): return get_base_dtype(arr_type.value_type ) else: return arr_type def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" if isinstance(lst[0] , __lowerCAmelCase ): change_first_primitive_element_in_list(lst[0] , __lowerCAmelCase ) else: snake_case_ : List[Any] = value @pytest.mark.parametrize('''optimized_int_type, expected_dtype''' , [(None, pa.intaa()), (Value('''int32''' ), pa.intaa())] ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> int: """simple docstring""" snake_case_ : Tuple = pa.array(TypedSequence(__lowerCAmelCase , optimized_int_type=__lowerCAmelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( '''col, expected_dtype''' , [ ('''attention_mask''', pa.inta()), ('''special_tokens_mask''', pa.inta()), ('''token_type_ids''', pa.inta()), ('''input_ids''', pa.intaa()), ('''other''', pa.intaa()), ] , ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Any: """simple docstring""" snake_case_ : Optional[int] = pa.array(OptimizedTypedSequence(__lowerCAmelCase , col=__lowerCAmelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications snake_case_ : Optional[Any] = copy.deepcopy(__lowerCAmelCase ) snake_case_ : Any = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(__lowerCAmelCase , __lowerCAmelCase ) snake_case_ : Union[str, Any] = pa.array(OptimizedTypedSequence(__lowerCAmelCase , col=__lowerCAmelCase ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize('''raise_exception''' , [False, True] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" snake_case_ : str = str(tmp_path / '''dataset-train.arrow''' ) try: with ArrowWriter(path=__lowerCAmelCase ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def lowerCamelCase_ ( _UpperCamelCase ) -> str: """simple docstring""" snake_case_ : Dict = '''mock://dataset-train.arrow''' with ArrowWriter(path=__lowerCAmelCase , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(__lowerCAmelCase ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) snake_case_ , snake_case_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(__lowerCAmelCase ) def lowerCamelCase_ ( ) -> Any: """simple docstring""" snake_case_ : Optional[Any] = pa.BufferOutputStream() with ParquetWriter(stream=__lowerCAmelCase ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) snake_case_ , snake_case_ : List[str] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 snake_case_ : Any = pa.BufferReader(output.getvalue() ) snake_case_ : int = pq.read_table(__lowerCAmelCase ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize('''embed_local_files''' , [False, True] ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> str: """simple docstring""" import PIL.Image snake_case_ : Union[str, Any] = str(tmp_path / '''test_image_rgb.jpg''' ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(__lowerCAmelCase , format='''png''' ) snake_case_ : Tuple = pa.BufferOutputStream() with ParquetWriter( stream=__lowerCAmelCase , features=Features({'''image''': Image()} ) , embed_local_files=__lowerCAmelCase ) as writer: writer.write({'''image''': image_path} ) writer.finalize() snake_case_ : Any = pa.BufferReader(output.getvalue() ) snake_case_ : Union[str, Any] = pq.read_table(__lowerCAmelCase ) snake_case_ : Optional[int] = pa_table.to_pydict() if embed_local_files: assert isinstance(out['''image'''][0]['''path'''] , __lowerCAmelCase ) with open(__lowerCAmelCase , '''rb''' ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def lowerCamelCase_ ( ) -> Optional[Any]: """simple docstring""" snake_case_ : Dict = pa.schema([pa.field('''col_1''' , pa.string() , nullable=__lowerCAmelCase )] ) snake_case_ : Optional[Any] = pa.BufferOutputStream() with ArrowWriter(stream=__lowerCAmelCase ) as writer: writer._build_writer(inferred_schema=__lowerCAmelCase ) assert writer._schema == pa.schema([pa.field('''col_1''' , pa.string() )] )
279
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase , cache_dir=UpperCAmelCase ) _UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(UpperCAmelCase , os.listdir(UpperCAmelCase )[0] , 'snapshots' ) )] _UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 4 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1e-3 assert np.abs(np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5e-1 _UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(UpperCAmelCase ) == num_samples def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , set_alpha_to_one=UpperCAmelCase , steps_offset=1 , ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=UpperCAmelCase , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = scheduler.create_state() _UpperCAmelCase = scheduler_state _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.random.PRNGKey(0 ) _UpperCAmelCase = 50 _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) # shard inputs and rng _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = jax.random.split(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1e-3 assert np.abs((np.abs(UpperCAmelCase , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5e-1 def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _UpperCAmelCase = jax.device_count() _UpperCAmelCase = num_samples * [prompt] _UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # With memory efficient attention _UpperCAmelCase , _UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=UpperCAmelCase , use_memory_efficient_attention=UpperCAmelCase , ) _UpperCAmelCase = replicate(UpperCAmelCase ) _UpperCAmelCase = pipeline.prepare_inputs(UpperCAmelCase ) _UpperCAmelCase = shard(UpperCAmelCase ) _UpperCAmelCase = pipeline(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , jit=UpperCAmelCase ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _UpperCAmelCase = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
39
0
import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): lowercase__ =yaml.safe_load( '\\nname: ""\nallow_empty: false\nallow_empty_text: true\nsubsections:\n - name: "Dataset Card for X" # First-level markdown heading\n allow_empty: false\n allow_empty_text: true\n subsections:\n - name: "Table of Contents"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Dataset Description"\n allow_empty: false\n allow_empty_text: false\n subsections:\n - name: "Dataset Summary"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Supported Tasks and Leaderboards"\n allow_empty: true\n allow_empty_text: true\n subsections: null\n - name: Languages\n allow_empty: false\n allow_empty_text: true\n subsections: null\n' ) lowercase__ ={ 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n#### Extra Ignored Subsection\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ={ 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Extra Ignored Subsection', 'text': '', 'is_empty_text': True, 'subsections': [], } ], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } lowercase__ ='\\n---\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ =( 'The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.' ) lowercase__ ='\\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ =( 'The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.' ) lowercase__ ='\\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found \'None\'.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n# Dataset Card My Dataset\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.' lowercase__ ='' lowercase__ ='The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.' lowercase__ ='\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' lowercase__ ='The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.' @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] ): assert ReadMe.from_string(__lowerCAmelCase , __lowerCAmelCase ).to_dict() == expected_dict @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] ): with pytest.raises(__lowerCAmelCase , match=re.escape(expected_error.format(path='''root''' ) ) ): __a : Union[str, Any] = ReadMe.from_string(__lowerCAmelCase , __lowerCAmelCase ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Tuple ): with pytest.raises(__lowerCAmelCase , match=re.escape(expected_error.format(path='''root''' ) ) ): ReadMe.from_string(__lowerCAmelCase , __lowerCAmelCase ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : Tuple ): ReadMe.from_string(__lowerCAmelCase , __lowerCAmelCase , suppress_parsing_errors=__lowerCAmelCase ) @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : int , lowerCAmelCase__ : Optional[Any] ): with tempfile.TemporaryDirectory() as tmp_dir: __a : Dict = Path(__lowerCAmelCase ) / '''README.md''' with open(__lowerCAmelCase , '''w+''' ) as readme_file: readme_file.write(__lowerCAmelCase ) __a : Tuple = ReadMe.from_readme(__lowerCAmelCase , __lowerCAmelCase ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[str] ): with tempfile.TemporaryDirectory() as tmp_dir: __a : Any = Path(__lowerCAmelCase ) / '''README.md''' with open(__lowerCAmelCase , '''w+''' ) as readme_file: readme_file.write(__lowerCAmelCase ) __a : Optional[Any] = expected_error.format(path=__lowerCAmelCase ) with pytest.raises(__lowerCAmelCase , match=re.escape(__lowerCAmelCase ) ): __a : Optional[Any] = ReadMe.from_readme(__lowerCAmelCase , __lowerCAmelCase ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmp_dir: __a : Any = Path(__lowerCAmelCase ) / '''README.md''' with open(__lowerCAmelCase , '''w+''' ) as readme_file: readme_file.write(__lowerCAmelCase ) __a : int = expected_error.format(path=__lowerCAmelCase ) with pytest.raises(__lowerCAmelCase , match=re.escape(__lowerCAmelCase ) ): ReadMe.from_readme(__lowerCAmelCase , __lowerCAmelCase ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] ): with tempfile.TemporaryDirectory() as tmp_dir: __a : Union[str, Any] = Path(__lowerCAmelCase ) / '''README.md''' with open(__lowerCAmelCase , '''w+''' ) as readme_file: readme_file.write(__lowerCAmelCase ) ReadMe.from_readme(__lowerCAmelCase , __lowerCAmelCase , suppress_parsing_errors=__lowerCAmelCase )
216
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = AlbertTokenizer UpperCamelCase__ = AlbertTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = True def UpperCamelCase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 'this is a test' _UpperCAmelCase = 'this is a test' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = '<pad>' _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(UpperCAmelCase ) , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 3_0000 ) def UpperCamelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = 'I was born in 92000, and this is falsé.' _UpperCAmelCase = tokenizer.tokenize(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(UpperCAmelCase ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [48, 25, 21, 1289] ) _UpperCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) _UpperCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase ) _UpperCAmelCase = tokenizer.encode('sequence builders' ) _UpperCAmelCase = tokenizer.encode('multi-sequence build' ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
39
0
from ... import PretrainedConfig _lowerCAmelCase : Any = { "sijunhe/nezha-cn-base": "https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json", } class __magic_name__ ( snake_case__ ): SCREAMING_SNAKE_CASE = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP SCREAMING_SNAKE_CASE = 'nezha' def __init__( self , __snake_case=2_1128 , __snake_case=768 , __snake_case=12 , __snake_case=12 , __snake_case=3072 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=512 , __snake_case=64 , __snake_case=2 , __snake_case=0.02 , __snake_case=1e-12 , __snake_case=0.1 , __snake_case=0 , __snake_case=2 , __snake_case=3 , __snake_case=True , **__snake_case , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , **__snake_case ) __a =vocab_size __a =hidden_size __a =num_hidden_layers __a =num_attention_heads __a =hidden_act __a =intermediate_size __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =max_position_embeddings __a =max_relative_position __a =type_vocab_size __a =initializer_range __a =layer_norm_eps __a =classifier_dropout __a =use_cache
218
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( snake_case__ , unittest.TestCase ): lowerCamelCase__: Tuple = TransfoXLTokenizer lowerCamelCase__: List[Any] = False lowerCamelCase__: str = False def _lowerCamelCase ( self: int ) -> Union[str, Any]: super().setUp() __UpperCAmelCase : str = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] __UpperCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def _lowerCamelCase ( self: int , **__lowerCamelCase: Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase : List[Any] = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowerCamelCase ( self: Any , __lowerCamelCase: Dict ) -> Tuple: __UpperCAmelCase : List[Any] = "<unk> UNwanted , running" __UpperCAmelCase : Tuple = "<unk> unwanted, running" return input_text, output_text def _lowerCamelCase ( self: str ) -> Tuple: __UpperCAmelCase : Optional[int] = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=__lowerCamelCase ) __UpperCAmelCase : Any = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowerCamelCase , ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [0, 4, 8, 7] ) def _lowerCamelCase ( self: int ) -> str: __UpperCAmelCase : Optional[int] = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ) , ["hello", "!", "how", "are", "you", "?"] ) def _lowerCamelCase ( self: List[Any] ) -> Union[str, Any]: __UpperCAmelCase : str = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def _lowerCamelCase ( self: List[str] ) -> Dict: __UpperCAmelCase : Any = TransfoXLTokenizer(lower_case=__lowerCamelCase ) __UpperCAmelCase : int = "Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?" __UpperCAmelCase : Tuple = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "\'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "\'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowerCamelCase ) , __lowerCamelCase ) def _lowerCamelCase ( self: Optional[Any] ) -> Tuple: __UpperCAmelCase : str = self.get_tokenizer() __UpperCAmelCase : Optional[Any] = len(__lowerCamelCase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowerCamelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , "new1" )
157
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "distilbert" UpperCamelCase__ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , UpperCAmelCase=3_0522 , UpperCAmelCase=512 , UpperCAmelCase=False , UpperCAmelCase=6 , UpperCAmelCase=12 , UpperCAmelCase=768 , UpperCAmelCase=4 * 768 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=0.02 , UpperCAmelCase=0.1 , UpperCAmelCase=0.2 , UpperCAmelCase=0 , **UpperCAmelCase , ): """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = sinusoidal_pos_embds _UpperCAmelCase = n_layers _UpperCAmelCase = n_heads _UpperCAmelCase = dim _UpperCAmelCase = hidden_dim _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation _UpperCAmelCase = initializer_range _UpperCAmelCase = qa_dropout _UpperCAmelCase = seq_classif_dropout super().__init__(**UpperCAmelCase , pad_token_id=UpperCAmelCase ) class __lowerCamelCase ( snake_case__): """simple docstring""" @property def UpperCamelCase ( self ): """simple docstring""" if self.task == "multiple-choice": _UpperCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _UpperCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
39
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _SCREAMING_SNAKE_CASE = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
158
import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() _a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCamelCase ( snake_case__): """simple docstring""" def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" os.makedirs(UpperCAmelCase , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {'source': 'What is love ?', 'target': 'life'} _UpperCAmelCase = {'train': 12, 'val': 2, 'test': 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: _UpperCAmelCase = '\n'.join([contents[field]] * n_lines[split] ) with open(os.path.join(UpperCAmelCase , F"""{split}.{field}""" ) , 'w' ) as f: f.write(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase = "pytorch" ): """simple docstring""" _UpperCAmelCase = self.get_auto_remove_tmp_dir() _UpperCAmelCase = os.path.join(UpperCAmelCase , 'output' ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'data' ) self._create_dummy_data(data_dir=UpperCAmelCase ) _UpperCAmelCase = F""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(F"""--gpus={gpus}""" ) if is_apex_available(): testargs.append('--fp16' ) else: testargs.append('--gpus=0' ) testargs.append('--distributed_backend=ddp_cpu' ) testargs.append('--num_processes=2' ) _UpperCAmelCase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(UpperCAmelCase , env=self.get_env() ) _UpperCAmelCase = os.path.join(UpperCAmelCase , 'metrics.json' ) with open(UpperCAmelCase ) as f: _UpperCAmelCase = json.load(UpperCAmelCase ) return result @require_torch_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 ) @require_torch_multi_gpu @require_ray def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self._run_finetune(gpus=1 , distributed_retriever='ray' ) self.assertGreaterEqual(result['test'][0]['test_avg_em'] , 0.2 )
39
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = """▁""" lowerCamelCase__ = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} lowerCamelCase__ = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } lowerCamelCase__ = {"""vinai/bartpho-syllable""": 1024} class A__ ( snake_case__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , a : List[Any] , a : Any , a : Union[str, Any]="<s>" , a : int="</s>" , a : int="</s>" , a : Tuple="<s>" , a : int="<unk>" , a : List[Any]="<pad>" , a : int="<mask>" , a : List[Any] = None , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token lowerCAmelCase__ : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , sp_model_kwargs=self.sp_model_kwargs , **a , ) lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Optional[int] = monolingual_vocab_file lowerCAmelCase__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility lowerCAmelCase__ : Union[str, Any] = {} lowerCAmelCase__ : Tuple = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(a ) not in self.fairseq_tokens_to_ids: lowerCAmelCase__ : str = cnt cnt += 1 with open(a , 'r' , encoding='utf-8' ) as f: for line in f.readlines(): lowerCAmelCase__ : Dict = line.strip().split()[0] lowerCAmelCase__ : List[Any] = len(self.fairseq_tokens_to_ids ) if str(a ) not in self.fairseq_tokens_to_ids: lowerCAmelCase__ : Optional[Any] = len(self.fairseq_tokens_to_ids ) lowerCAmelCase__ : str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = self.__dict__.copy() lowerCAmelCase__ : Optional[Any] = None lowerCAmelCase__ : Dict = self.sp_model.serialized_model_proto() return state def __setstate__( self : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : str = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase__ : List[Any] = {} lowerCAmelCase__ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _lowerCamelCase ( self : int , a : Dict , a : Tuple = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase__ : Tuple = [self.cls_token_id] lowerCAmelCase__ : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowerCamelCase ( self : Dict , a : Any , a : Tuple = None , a : str = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : Dict , a : Dict , a : List[str] = None ): '''simple docstring''' lowerCAmelCase__ : Any = [self.sep_token_id] lowerCAmelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _lowerCamelCase ( self : Any ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowerCamelCase ( self : List[str] , a : List[Any] ): '''simple docstring''' return self.sp_model.encode(a , out_type=a ) def _lowerCamelCase ( self : List[str] , a : Any ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def _lowerCamelCase ( self : Optional[int] , a : List[str] ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def _lowerCamelCase ( self : int , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Dict = ''.join(a ).replace(a , ' ' ).strip() return out_string def _lowerCamelCase ( self : Tuple , a : Any , a : str = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase__ : Optional[int] = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a ) elif not os.path.isfile(self.vocab_file ): with open(a , 'wb' ) as fi: lowerCAmelCase__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(a , 'w' , encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f'''{str(a )} \n''' ) return out_vocab_file, out_monolingual_vocab_file
212
class __lowerCamelCase : """simple docstring""" def __init__( self ): """simple docstring""" _UpperCAmelCase = {} # Mapping from char to TrieNode _UpperCAmelCase = False def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: _UpperCAmelCase = TrieNode() _UpperCAmelCase = curr.nodes[char] _UpperCAmelCase = True def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self for char in word: if char not in curr.nodes: return False _UpperCAmelCase = curr.nodes[char] return curr.is_leaf def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" def _delete(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> bool: if index == len(UpperCAmelCase ): # If word does not exist if not curr.is_leaf: return False _UpperCAmelCase = False return len(curr.nodes ) == 0 _UpperCAmelCase = word[index] _UpperCAmelCase = curr.nodes.get(UpperCAmelCase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted _UpperCAmelCase = _delete(UpperCAmelCase , UpperCAmelCase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCAmelCase , 0 ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" if node.is_leaf: print(__lowerCAmelCase , end=' ' ) for key, value in node.nodes.items(): print_words(__lowerCAmelCase , word + key ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = TrieNode() root.insert_many(__lowerCAmelCase ) # print_words(root, "") assert all(root.find(__lowerCAmelCase ) for word in words ) assert root.find('banana' ) assert not root.find('bandanas' ) assert not root.find('apps' ) assert root.find('apple' ) assert root.find('all' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" print(str(__lowerCAmelCase ) , 'works!' if passes else 'doesn\'t work :(' ) def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" print_results('Testing trie functionality' , test_trie() ) if __name__ == "__main__": main()
39
0
import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertModel, ) class UpperCAmelCase : '''simple docstring''' def __init__( self : Union[str, Any] ,A : Tuple ,A : Union[str, Any]=13 ,A : List[Any]=7 ,A : List[Any]=True ,A : List[str]=True ,A : List[str]=True ,A : Tuple=True ,A : Optional[Any]=99 ,A : Union[str, Any]=64 ,A : str=32 ,A : int=5 ,A : Union[str, Any]=4 ,A : Union[str, Any]=37 ,A : List[Any]="gelu" ,A : Any=0.1 ,A : Any=0.1 ,A : Any=5_12 ,A : Optional[Any]=16 ,A : int=2 ,A : Union[str, Any]=0.02 ,A : Any=3 ,A : Union[str, Any]=4 ,A : Union[str, Any]=None ,): __A = parent __A = batch_size __A = seq_length __A = is_training __A = use_input_mask __A = use_token_type_ids __A = use_labels __A = vocab_size __A = hidden_size __A = embedding_size __A = num_hidden_layers __A = num_attention_heads __A = intermediate_size __A = hidden_act __A = hidden_dropout_prob __A = attention_probs_dropout_prob __A = max_position_embeddings __A = type_vocab_size __A = type_sequence_label_size __A = initializer_range __A = num_labels __A = num_choices __A = scope def UpperCamelCase_ ( self : Optional[int] ): __A = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __A = None if self.use_input_mask: __A = random_attention_mask([self.batch_size, self.seq_length] ) __A = None if self.use_token_type_ids: __A = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __A = None __A = None __A = None if self.use_labels: __A = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __A = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __A = ids_tensor([self.batch_size] ,self.num_choices ) __A = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self : int ): return MobileBertConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,embedding_size=self.embedding_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A ,initializer_range=self.initializer_range ,) def UpperCamelCase_ ( self : Dict ,A : int ,A : Tuple ,A : Optional[int] ,A : List[Any] ,A : Optional[int] ,A : Optional[int] ,A : Optional[Any] ): __A = MobileBertModel(config=A ) model.to(A ) model.eval() __A = model(A ,attention_mask=A ,token_type_ids=A ) __A = model(A ,token_type_ids=A ) __A = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase_ ( self : Tuple ,A : Tuple ,A : Union[str, Any] ,A : Tuple ,A : Union[str, Any] ,A : Union[str, Any] ,A : int ,A : Optional[int] ): __A = MobileBertForMaskedLM(config=A ) model.to(A ) model.eval() __A = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self : Optional[int] ,A : str ,A : Any ,A : Optional[int] ,A : Optional[Any] ,A : List[Any] ,A : Tuple ,A : Dict ): __A = MobileBertForNextSentencePrediction(config=A ) model.to(A ) model.eval() __A = model( A ,attention_mask=A ,token_type_ids=A ,labels=A ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 2) ) def UpperCamelCase_ ( self : str ,A : Optional[Any] ,A : Optional[int] ,A : Optional[Any] ,A : Tuple ,A : List[str] ,A : Optional[int] ,A : Dict ): __A = MobileBertForPreTraining(config=A ) model.to(A ) model.eval() __A = model( A ,attention_mask=A ,token_type_ids=A ,labels=A ,next_sentence_label=A ,) self.parent.assertEqual(result.prediction_logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape ,(self.batch_size, 2) ) def UpperCamelCase_ ( self : Optional[Any] ,A : Any ,A : List[str] ,A : List[str] ,A : Optional[int] ,A : Dict ,A : Optional[int] ,A : Any ): __A = MobileBertForQuestionAnswering(config=A ) model.to(A ) model.eval() __A = model( A ,attention_mask=A ,token_type_ids=A ,start_positions=A ,end_positions=A ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self : str ,A : int ,A : Tuple ,A : Optional[Any] ,A : int ,A : Union[str, Any] ,A : List[Any] ,A : Dict ): __A = self.num_labels __A = MobileBertForSequenceClassification(A ) model.to(A ) model.eval() __A = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self : Optional[int] ,A : Optional[Any] ,A : List[str] ,A : List[str] ,A : Optional[int] ,A : int ,A : Any ,A : Any ): __A = self.num_labels __A = MobileBertForTokenClassification(config=A ) model.to(A ) model.eval() __A = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self : int ,A : Dict ,A : List[str] ,A : List[str] ,A : List[Any] ,A : Optional[Any] ,A : str ,A : Any ): __A = self.num_choices __A = MobileBertForMultipleChoice(config=A ) model.to(A ) model.eval() __A = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() __A = token_type_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() __A = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() __A = model( A ,attention_mask=A ,token_type_ids=A ,labels=A ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self : Any ): __A = self.prepare_config_and_inputs() ( ( __A ) , ( __A ) , ( __A ) , ( __A ) , ( __A ) , ( __A ) , ( __A ) , ) = config_and_inputs __A = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase ( snake_case__ , snake_case__ , unittest.TestCase ): '''simple docstring''' snake_case_ = ( ( MobileBertModel, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, ) if is_torch_available() else () ) snake_case_ = ( { "feature-extraction": MobileBertModel, "fill-mask": MobileBertForMaskedLM, "question-answering": MobileBertForQuestionAnswering, "text-classification": MobileBertForSequenceClassification, "token-classification": MobileBertForTokenClassification, "zero-shot": MobileBertForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = True def UpperCamelCase_ ( self : int ,A : Dict ,A : Optional[int] ,A : int=False ): __A = super()._prepare_for_class(A ,A ,return_labels=A ) if return_labels: if model_class in get_values(A ): __A = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) ,dtype=torch.long ,device=A ) __A = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A ) return inputs_dict def UpperCamelCase_ ( self : List[str] ): __A = MobileBertModelTester(self ) __A = ConfigTester(self ,config_class=A ,hidden_size=37 ) def UpperCamelCase_ ( self : List[str] ): self.config_tester.run_common_tests() def UpperCamelCase_ ( self : Optional[int] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*A ) def UpperCamelCase_ ( self : List[Any] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*A ) def UpperCamelCase_ ( self : int ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*A ) def UpperCamelCase_ ( self : List[Any] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*A ) def UpperCamelCase_ ( self : Any ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*A ) def UpperCamelCase_ ( self : Optional[Any] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*A ) def UpperCamelCase_ ( self : Dict ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*A ) def UpperCamelCase_ ( self : Tuple ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*A ) def UpperCAmelCase ( a_ ) -> Optional[Any]: """simple docstring""" return torch.tensor( __lowerCAmelCase , dtype=torch.long , device=__lowerCAmelCase , ) SCREAMING_SNAKE_CASE :Union[str, Any] = 1E-3 @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self : Optional[Any] ): __A = MobileBertModel.from_pretrained("google/mobilebert-uncased" ).to(A ) __A = _long_tensor([[1_01, 71_10, 10_05, 10_56, 20_23, 1_13_33, 1_74_13, 10_29, 1_02]] ) with torch.no_grad(): __A = model(A )[0] __A = torch.Size((1, 9, 5_12) ) self.assertEqual(output.shape ,A ) __A = torch.tensor( [ [ [-2.4736526E07, 8.2691656E04, 1.6521838E05], [-5.7541704E-01, 3.9056022E00, 4.4011507E00], [2.6047359E00, 1.5677652E00, -1.7324188E-01], ] ] ,device=A ,) # MobileBERT results range from 10e0 to 10e8. Even a 0.0000001% difference with a value of 10e8 results in a # ~1 difference, it's therefore not a good idea to measure using addition. # Here, we instead divide the expected result with the result in order to obtain ~1. We then check that the # result is held between bounds: 1 - TOLERANCE < expected_result / result < 1 + TOLERANCE __A = torch.all((expected_slice / output[..., :3, :3]) >= 1 - TOLERANCE ) __A = torch.all((expected_slice / output[..., :3, :3]) <= 1 + TOLERANCE ) self.assertTrue(lower_bound and upper_bound )
15
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. _a = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class __lowerCamelCase ( unittest.TestCase): """simple docstring""" UpperCamelCase__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = ZeroShotClassificationPipeline( model=UpperCAmelCase , tokenizer=UpperCAmelCase , candidate_labels=['polics', 'health'] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # No kwarg _UpperCAmelCase = classifier('Who are you voting for in 2020?' , ['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics'] ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels='politics, public health' ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier('Who are you voting for in 2020?' , candidate_labels=['politics', 'public health'] ) self.assertEqual( UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['scores'] ) ) , 1.0 ) _UpperCAmelCase = classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='This text is about {}' ) self.assertEqual(UpperCAmelCase , {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase )]} ) # https://github.com/huggingface/transformers/issues/13846 _UpperCAmelCase = classifier(['I am happy'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(1 ) ] , ) _UpperCAmelCase = classifier(['I am happy', 'I am sad'] , ['positive', 'negative'] ) self.assertEqual( UpperCAmelCase , [ {'sequence': ANY(UpperCAmelCase ), 'labels': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )], 'scores': [ANY(UpperCAmelCase ), ANY(UpperCAmelCase )]} for i in range(2 ) ] , ) with self.assertRaises(UpperCAmelCase ): classifier('' , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier(UpperCAmelCase , candidate_labels='politics' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels='' ) with self.assertRaises(UpperCAmelCase ): classifier('Who are you voting for in 2020?' , candidate_labels=UpperCAmelCase ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template='Not formatting template' , ) with self.assertRaises(UpperCAmelCase ): classifier( 'Who are you voting for in 2020?' , candidate_labels='politics' , hypothesis_template=UpperCAmelCase , ) self.run_entailment_id(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = zero_shot_classifier.model.config _UpperCAmelCase = config.labelaid _UpperCAmelCase = zero_shot_classifier.entailment_id _UpperCAmelCase = {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) _UpperCAmelCase = {'entailment': 0, 'neutral': 1, 'contradiction': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 0, 'NON-ENTAIL': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _UpperCAmelCase = {'ENTAIL': 2, 'NEUTRAL': 1, 'CONTR': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) _UpperCAmelCase = original_labelaid self.assertEqual(UpperCAmelCase , zero_shot_classifier.entailment_id ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( 'Who are you voting for in 2020?' * 100 , candidate_labels=['politics', 'public health', 'science'] ) @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='pt' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline( 'zero-shot-classification' , model='sshleifer/tiny-distilbert-base-cased-distilled-squad' , framework='tf' , ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['science', 'public health', 'politics'], 'scores': [0.3_33, 0.3_33, 0.3_33], } , ) @slow @require_torch def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='pt' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , ) @slow @require_tf def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = pipeline('zero-shot-classification' , model='roberta-large-mnli' , framework='tf' ) _UpperCAmelCase = zero_shot_classifier( 'Who are you voting for in 2020?' , candidate_labels=['politics', 'public health', 'science'] ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': 'Who are you voting for in 2020?', 'labels': ['politics', 'public health', 'science'], 'scores': [0.9_76, 0.0_15, 0.0_09], } , ) _UpperCAmelCase = zero_shot_classifier( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural networks' ' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder' ' through an attention mechanism. We propose a new simple network architecture, the Transformer, based' ' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two' ' machine translation tasks show these models to be superior in quality while being more parallelizable' ' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014' ' English-to-German translation task, improving over the existing best results, including ensembles by' ' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new' ' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small' ' fraction of the training costs of the best models from the literature. We show that the Transformer' ' generalizes well to other tasks by applying it successfully to English constituency parsing both with' ' large and limited training data.' , candidate_labels=['machine learning', 'statistics', 'translation', 'vision'] , multi_label=UpperCAmelCase , ) self.assertEqual( nested_simplify(UpperCAmelCase ) , { 'sequence': ( 'The dominant sequence transduction models are based on complex recurrent or convolutional neural' ' networks in an encoder-decoder configuration. The best performing models also connect the' ' encoder and decoder through an attention mechanism. We propose a new simple network' ' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence' ' and convolutions entirely. Experiments on two machine translation tasks show these models to be' ' superior in quality while being more parallelizable and requiring significantly less time to' ' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,' ' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014' ' English-to-French translation task, our model establishes a new single-model state-of-the-art' ' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training' ' costs of the best models from the literature. We show that the Transformer generalizes well to' ' other tasks by applying it successfully to English constituency parsing both with large and' ' limited training data.' ), 'labels': ['translation', 'machine learning', 'vision', 'statistics'], 'scores': [0.8_17, 0.7_13, 0.0_18, 0.0_18], } , )
39
0
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class __snake_case : def __init__( self : Tuple , _lowercase : Optional[Any] , _lowercase : Dict=13 , _lowercase : Optional[int]=7 , _lowercase : int=True , _lowercase : List[str]=True , _lowercase : str=True , _lowercase : Optional[Any]=True , _lowercase : int=99 , _lowercase : Tuple=32 , _lowercase : Dict=2 , _lowercase : Any=4 , _lowercase : List[Any]=37 , _lowercase : Any="gelu" , _lowercase : Any=0.1 , _lowercase : Tuple=0.1 , _lowercase : Optional[int]=5_12 , _lowercase : int=16 , _lowercase : str=2 , _lowercase : int=0.02 , _lowercase : Dict=3 , _lowercase : Any=4 , _lowercase : List[Any]=None , _lowercase : Optional[Any]=0 , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = type_sequence_label_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = num_choices SCREAMING_SNAKE_CASE__ = scope SCREAMING_SNAKE_CASE__ = projection_dim def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE__ = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.num_choices ) SCREAMING_SNAKE_CASE__ = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) SCREAMING_SNAKE_CASE__ = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __a ( self : Any , _lowercase : str , _lowercase : Optional[int] , _lowercase : List[str] , _lowercase : Optional[int] , _lowercase : Union[str, Any] , _lowercase : List[str] , _lowercase : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFDPRContextEncoder(config=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase , token_type_ids=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def __a ( self : int , _lowercase : Dict , _lowercase : Any , _lowercase : List[Any] , _lowercase : List[str] , _lowercase : List[str] , _lowercase : Optional[int] , _lowercase : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFDPRQuestionEncoder(config=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase , token_type_ids=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def __a ( self : Any , _lowercase : Any , _lowercase : Union[str, Any] , _lowercase : str , _lowercase : int , _lowercase : Optional[Any] , _lowercase : str , _lowercase : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFDPRReader(config=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""input_ids""": input_ids} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): lowerCAmelCase_ = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) lowerCAmelCase_ = {"feature-extraction": TFDPRQuestionEncoder} if is_tf_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFDPRModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=_lowercase , hidden_size=37 ) def __a ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*_lowercase ) def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*_lowercase ) def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*_lowercase ) @slow def __a ( self : Any ): """simple docstring""" for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = TFDPRContextEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = TFDPRContextEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = TFDPRQuestionEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = TFDPRReader.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) @require_tf class __snake_case ( unittest.TestCase ): @slow def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFDPRQuestionEncoder.from_pretrained("""facebook/dpr-question_encoder-single-nq-base""" ) SCREAMING_SNAKE_CASE__ = tf.constant( [[1_01, 75_92, 10_10, 20_03, 20_26, 38_99, 1_01_40, 10_29, 1_02]] ) # [CLS] hello, is my dog cute? [SEP] SCREAMING_SNAKE_CASE__ = model(_lowercase )[0] # embedding shape = (1, 768) # compare the actual values for a slice. SCREAMING_SNAKE_CASE__ = tf.constant( [ [ 0.03_23_62_53, 0.12_75_33_35, 0.16_81_85_09, 0.00_27_97_86, 0.3_89_69_33, 0.24_26_49_45, 0.2_17_89_71, -0.02_33_52_27, -0.08_48_19_59, -0.14_32_41_17, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1E-4 ) )
219
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _a = get_logger(__name__) class __lowerCamelCase ( enum.Enum): """simple docstring""" UpperCamelCase__ = "all_checks" UpperCamelCase__ = "basic_checks" UpperCamelCase__ = "no_checks" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None )-> str: """simple docstring""" if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] _UpperCAmelCase = ' for ' + verification_name if verification_name is not None else '' if len(__lowerCAmelCase ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" class __lowerCamelCase ( snake_case__): """simple docstring""" def __A ( __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise ExpectedMoreSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) if len(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) > 0: raise UnexpectedSplits(str(set(__lowerCAmelCase ) - set(__lowerCAmelCase ) ) ) _UpperCAmelCase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__lowerCAmelCase ) > 0: raise NonMatchingSplitsSizesError(str(__lowerCAmelCase ) ) logger.info('All the splits matched successfully.' ) def __A ( __lowerCAmelCase , __lowerCAmelCase = True )-> dict: """simple docstring""" if record_checksum: _UpperCAmelCase = shaaaa() with open(__lowerCAmelCase , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__lowerCAmelCase ) _UpperCAmelCase = m.hexdigest() else: _UpperCAmelCase = None return {"num_bytes": os.path.getsize(__lowerCAmelCase ), "checksum": checksum} def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
39
0
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" __lowerCamelCase = StableDiffusionDiffEditPipeline __lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'height', 'width', 'image'} | {'image_latents'} __lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {'image'} | {'image_latents'} __lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __lowerCamelCase = frozenset([] ) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0 ) A__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowercase , ) A__ = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowercase , set_alpha_to_one=lowercase , ) A__ = DDIMInverseScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowercase , set_alpha_to_zero=lowercase , ) torch.manual_seed(0 ) A__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) A__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) A__ = CLIPTextModel(lowercase ) A__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) A__ = { "unet": unet, "scheduler": scheduler, "inverse_scheduler": inverse_scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase ( self , lowercase , lowercase=0 ) -> int: '''simple docstring''' A__ = floats_tensor((1, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) A__ = floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) if str(lowercase ).startswith("mps" ): A__ = torch.manual_seed(lowercase ) else: A__ = torch.Generator(device=lowercase ).manual_seed(lowercase ) A__ = { "prompt": "a dog and a newt", "mask_image": mask, "image_latents": latents, "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase ( self , lowercase , lowercase=0 ) -> Dict: '''simple docstring''' A__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) A__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] A__ = Image.fromarray(np.uinta(lowercase ) ).convert("RGB" ) if str(lowercase ).startswith("mps" ): A__ = torch.manual_seed(lowercase ) else: A__ = torch.Generator(device=lowercase ).manual_seed(lowercase ) A__ = { "image": image, "source_prompt": "a cat and a frog", "target_prompt": "a dog and a newt", "generator": generator, "num_inference_steps": 2, "num_maps_per_mask": 2, "mask_encode_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase ( self , lowercase , lowercase=0 ) -> Dict: '''simple docstring''' A__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) A__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] A__ = Image.fromarray(np.uinta(lowercase ) ).convert("RGB" ) if str(lowercase ).startswith("mps" ): A__ = torch.manual_seed(lowercase ) else: A__ = torch.Generator(device=lowercase ).manual_seed(lowercase ) A__ = { "image": image, "prompt": "a cat and a frog", "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "decode_latents": True, "output_type": "numpy", } return inputs def UpperCamelCase ( self ) -> int: '''simple docstring''' if not hasattr(self.pipeline_class , "_optional_components" ): return A__ = self.get_dummy_components() A__ = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowercase , lowercase , lowercase ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) A__ = self.get_dummy_inputs(lowercase ) A__ = pipe(**lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowercase ) A__ = self.pipeline_class.from_pretrained(lowercase ) pipe_loaded.to(lowercase ) pipe_loaded.set_progress_bar_config(disable=lowercase ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowercase , lowercase ) is None , F'`{optional_component}` did not stay set to None after loading.' , ) A__ = self.get_dummy_inputs(lowercase ) A__ = pipe_loaded(**lowercase )[0] A__ = np.abs(output - output_loaded ).max() self.assertLess(lowercase , 1e-4 ) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' A__ = "cpu" A__ = self.get_dummy_components() A__ = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) A__ = self.get_dummy_mask_inputs(lowercase ) A__ = pipe.generate_mask(**lowercase ) A__ = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) A__ = np.array([0] * 9 ) A__ = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' A__ = "cpu" A__ = self.get_dummy_components() A__ = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) A__ = self.get_dummy_inversion_inputs(lowercase ) A__ = pipe.invert(**lowercase ).images A__ = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) A__ = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) A__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def UpperCamelCase ( self ) -> int: '''simple docstring''' A__ = "cpu" A__ = self.get_dummy_components() A__ = {"beta_start": 0.0_0085, "beta_end": 0.012, "beta_schedule": "scaled_linear"} A__ = DPMSolverMultistepScheduler(**lowercase ) A__ = DPMSolverMultistepInverseScheduler(**lowercase ) A__ = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) A__ = self.get_dummy_inversion_inputs(lowercase ) A__ = pipe.invert(**lowercase ).images A__ = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) A__ = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.5_1050, 0.5015, 0.4407, 0.4799] , ) A__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1e-3 ) @require_torch_gpu @slow class a__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def UpperCamelCase ( cls ) -> List[str]: '''simple docstring''' A__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png" ) A__ = raw_image.convert("RGB" ).resize((768, 768) ) A__ = raw_image def UpperCamelCase ( self ) -> List[Any]: '''simple docstring''' A__ = torch.manual_seed(0 ) A__ = StableDiffusionDiffEditPipeline.from_pretrained( "stabilityai/stable-diffusion-2-1" , safety_checker=lowercase , torch_dtype=torch.floataa ) A__ = DDIMScheduler.from_config(pipe.scheduler.config ) A__ = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowercase ) A__ = "a bowl of fruit" A__ = "a bowl of pears" A__ = pipe.generate_mask( image=self.raw_image , source_prompt=lowercase , target_prompt=lowercase , generator=lowercase , ) A__ = pipe.invert( prompt=lowercase , image=self.raw_image , inpaint_strength=0.7 , generator=lowercase ).latents A__ = pipe( prompt=lowercase , mask_image=lowercase , image_latents=lowercase , generator=lowercase , negative_prompt=lowercase , inpaint_strength=0.7 , output_type="numpy" , ).images[0] A__ = ( np.array( load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/diffedit/pears.png" ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1 def UpperCamelCase ( self ) -> List[Any]: '''simple docstring''' A__ = torch.manual_seed(0 ) A__ = StableDiffusionDiffEditPipeline.from_pretrained( "stabilityai/stable-diffusion-2-1" , safety_checker=lowercase , torch_dtype=torch.floataa ) A__ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) A__ = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowercase ) A__ = "a bowl of fruit" A__ = "a bowl of pears" A__ = pipe.generate_mask( image=self.raw_image , source_prompt=lowercase , target_prompt=lowercase , generator=lowercase , ) A__ = pipe.invert( prompt=lowercase , image=self.raw_image , inpaint_strength=0.7 , generator=lowercase , num_inference_steps=25 , ).latents A__ = pipe( prompt=lowercase , mask_image=lowercase , image_latents=lowercase , generator=lowercase , negative_prompt=lowercase , inpaint_strength=0.7 , num_inference_steps=25 , output_type="numpy" , ).images[0] A__ = ( np.array( load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/diffedit/pears.png" ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5e-1
68
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=16 , UpperCAmelCase=[1, 2, 1] , UpperCAmelCase=[2, 2, 4] , UpperCAmelCase=2 , UpperCAmelCase=2.0 , UpperCAmelCase=True , UpperCAmelCase=0.0 , UpperCAmelCase=0.0 , UpperCAmelCase=0.1 , UpperCAmelCase="gelu" , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=0.02 , UpperCAmelCase=1e-5 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=10 , UpperCAmelCase=8 , UpperCAmelCase=["stage1", "stage2", "stage3"] , UpperCAmelCase=[1, 2, 3] , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = embed_dim _UpperCAmelCase = depths _UpperCAmelCase = num_heads _UpperCAmelCase = window_size _UpperCAmelCase = mlp_ratio _UpperCAmelCase = qkv_bias _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = drop_path_rate _UpperCAmelCase = hidden_act _UpperCAmelCase = use_absolute_embeddings _UpperCAmelCase = patch_norm _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = initializer_range _UpperCAmelCase = is_training _UpperCAmelCase = scope _UpperCAmelCase = use_labels _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = encoder_stride _UpperCAmelCase = out_features _UpperCAmelCase = out_indices def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) _UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = model(UpperCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(UpperCAmelCase ): _UpperCAmelCase = ['stem'] _UpperCAmelCase = MaskFormerSwinBackbone(config=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) UpperCamelCase__ = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) # Swin has a different seq_length _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = 3 _UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True self.check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(UpperCAmelCase ): _UpperCAmelCase = 0 return t def check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase={} ): with torch.no_grad(): _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = model(**UpperCAmelCase , return_dict=UpperCAmelCase , **UpperCAmelCase ).to_tuple() def recursive_check(UpperCAmelCase , UpperCAmelCase ): if isinstance(UpperCAmelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(UpperCAmelCase , UpperCAmelCase ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(UpperCAmelCase , UpperCAmelCase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(UpperCAmelCase ) , set_nan_tensor_to_zero(UpperCAmelCase ) , atol=1e-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}. Dict has""" F""" `nan`: {torch.isnan(UpperCAmelCase ).any()} and `inf`: {torch.isinf(UpperCAmelCase )}.""" ) , ) recursive_check(UpperCAmelCase , UpperCAmelCase ) for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) _UpperCAmelCase = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) check_equivalence(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , {'output_hidden_states': True} ) @require_torch class __lowerCamelCase ( unittest.TestCase , snake_case__): """simple docstring""" UpperCamelCase__ = (MaskFormerSwinBackbone,) if is_torch_available() else () UpperCamelCase__ = MaskFormerSwinConfig def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = MaskFormerSwinModelTester(self ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: _UpperCAmelCase = backbone_class(UpperCAmelCase ) backbone.to(UpperCAmelCase ) backbone.eval() _UpperCAmelCase = backbone(**UpperCAmelCase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , UpperCAmelCase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True _UpperCAmelCase = backbone(**UpperCAmelCase , output_hidden_states=UpperCAmelCase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: _UpperCAmelCase = backbone(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertIsNotNone(outputs.attentions )
39
0
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel UpperCAmelCase__ : List[str] = { 'text_branch': 'text_model', 'audio_branch': 'audio_model.audio_encoder', 'attn': 'attention.self', 'self.proj': 'output.dense', 'attention.self_mask': 'attn_mask', 'mlp.fc1': 'intermediate.dense', 'mlp.fc2': 'output.dense', 'norm1': 'layernorm_before', 'norm2': 'layernorm_after', 'bn0': 'batch_norm', } UpperCAmelCase__ : Optional[Any] = AutoFeatureExtractor.from_pretrained('laion/clap-htsat-unfused', truncation='rand_trunc') def lowerCamelCase__ ( a , a=False ) -> Optional[Any]: _A , _A: Dict = create_model( '''HTSAT-tiny''' , '''roberta''' , __lowerCAmelCase , precision='''fp32''' , device='''cuda:0''' if torch.cuda.is_available() else '''cpu''' , enable_fusion=__lowerCAmelCase , fusion_type='''aff_2d''' if enable_fusion else None , ) return model, model_cfg def lowerCamelCase__ ( a ) -> List[str]: _A: str = {} _A: List[str] = R'''.*sequential.(\d+).*''' _A: Any = R'''.*_projection.(\d+).*''' for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _A: Any = key.replace(__lowerCAmelCase , __lowerCAmelCase ) if re.match(__lowerCAmelCase , __lowerCAmelCase ): # replace sequential layers with list _A: Union[str, Any] = re.match(__lowerCAmelCase , __lowerCAmelCase ).group(1 ) _A: List[str] = key.replace(f"""sequential.{sequential_layer}.""" , f"""layers.{int(__lowerCAmelCase )//3}.linear.""" ) elif re.match(__lowerCAmelCase , __lowerCAmelCase ): _A: int = int(re.match(__lowerCAmelCase , __lowerCAmelCase ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... _A: Dict = 1 if projecton_layer == 0 else 2 _A: Any = key.replace(f"""_projection.{projecton_layer}.""" , f"""_projection.linear{transformers_projection_layer}.""" ) if "audio" and "qkv" in key: # split qkv into query key and value _A: Union[str, Any] = value _A: Dict = mixed_qkv.size(0 ) // 3 _A: Optional[Any] = mixed_qkv[:qkv_dim] _A: Tuple = mixed_qkv[qkv_dim : qkv_dim * 2] _A: str = mixed_qkv[qkv_dim * 2 :] _A: List[Any] = query_layer _A: List[str] = key_layer _A: Tuple = value_layer else: _A: Optional[Any] = value return model_state_dict def lowerCamelCase__ ( a , a , a , a=False ) -> int: _A , _A: Optional[Any] = init_clap(__lowerCAmelCase , enable_fusion=__lowerCAmelCase ) clap_model.eval() _A: Union[str, Any] = clap_model.state_dict() _A: Dict = rename_state_dict(__lowerCAmelCase ) _A: List[Any] = ClapConfig() _A: Dict = enable_fusion _A: List[str] = ClapModel(__lowerCAmelCase ) # ignore the spectrogram embedding layer model.load_state_dict(__lowerCAmelCase , strict=__lowerCAmelCase ) model.save_pretrained(__lowerCAmelCase ) transformers_config.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ : Any = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument('--enable_fusion', action='store_true', help='Whether to enable fusion or not') UpperCAmelCase__ : Tuple = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
121
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __lowerCamelCase ( snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = TransfoXLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" super().setUp() _UpperCAmelCase = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def UpperCamelCase ( self , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = '<unk> UNwanted , running' _UpperCAmelCase = '<unk> unwanted, running' return input_text, output_text def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=UpperCAmelCase ) _UpperCAmelCase = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(UpperCAmelCase , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [0, 4, 8, 7] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TransfoXLTokenizer(lower_case=UpperCAmelCase ) _UpperCAmelCase = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' _UpperCAmelCase = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(UpperCAmelCase ) , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(UpperCAmelCase ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(UpperCAmelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
39
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class _UpperCAmelCase( snake_case__ ): lowercase__ = 'lilt' def __init__( self , __a=3_05_22 , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=2 , __a=0.02 , __a=1e-12 , __a=0 , __a="absolute" , __a=None , __a=4 , __a=10_24 , **__a , ) -> int: '''simple docstring''' super().__init__(pad_token_id=__a , **__a) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = position_embedding_type _UpperCamelCase = classifier_dropout _UpperCamelCase = channel_shrink_ratio _UpperCamelCase = max_ad_position_embeddings
194
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = {'''configuration_opt''': ['''OPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OPTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''OPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OPTForCausalLM''', '''OPTModel''', '''OPTPreTrainedModel''', '''OPTForSequenceClassification''', '''OPTForQuestionAnswering''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['''TFOPTForCausalLM''', '''TFOPTModel''', '''TFOPTPreTrainedModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ '''FlaxOPTForCausalLM''', '''FlaxOPTModel''', '''FlaxOPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
39
0
import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Optional[int] = StableDiffusionKDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' ) snake_case_ : int = sd_pipe.to(__magic_name__ ) sd_pipe.set_progress_bar_config(disable=__magic_name__ ) sd_pipe.set_scheduler('''sample_euler''' ) snake_case_ : int = '''A painting of a squirrel eating a burger''' snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[Any] = sd_pipe([prompt] , generator=__magic_name__ , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''' ) snake_case_ : Optional[Any] = output.images snake_case_ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case_ : List[Any] = np.array([0.0_447, 0.0_492, 0.0_468, 0.0_408, 0.0_383, 0.0_408, 0.0_354, 0.0_380, 0.0_339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Tuple = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) snake_case_ : List[str] = sd_pipe.to(__magic_name__ ) sd_pipe.set_progress_bar_config(disable=__magic_name__ ) sd_pipe.set_scheduler('''sample_euler''' ) snake_case_ : List[str] = '''A painting of a squirrel eating a burger''' snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Optional[int] = sd_pipe([prompt] , generator=__magic_name__ , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''' ) snake_case_ : Any = output.images snake_case_ : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case_ : Optional[int] = np.array([0.1_237, 0.1_320, 0.1_438, 0.1_359, 0.1_390, 0.1_132, 0.1_277, 0.1_175, 0.1_112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-1 def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) snake_case_ : Union[str, Any] = sd_pipe.to(__magic_name__ ) sd_pipe.set_progress_bar_config(disable=__magic_name__ ) sd_pipe.set_scheduler('''sample_dpmpp_2m''' ) snake_case_ : List[Any] = '''A painting of a squirrel eating a burger''' snake_case_ : Optional[int] = torch.manual_seed(0 ) snake_case_ : List[str] = sd_pipe( [prompt] , generator=__magic_name__ , guidance_scale=7.5 , num_inference_steps=15 , output_type='''np''' , use_karras_sigmas=__magic_name__ , ) snake_case_ : Optional[int] = output.images snake_case_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) snake_case_ : Any = np.array( [0.11_381_689, 0.12_112_921, 0.1_389_457, 0.12_549_606, 0.1_244_964, 0.10_831_517, 0.11_562_866, 0.10_867_816, 0.10_499_048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
279
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> None: """simple docstring""" _UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ), F"""{len(__lowerCAmelCase )} != {len(__lowerCAmelCase )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) _a = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _a = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def __A ( __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" try: _UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(__lowerCAmelCase ) ) def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(__lowerCAmelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def __A ( __lowerCAmelCase , __lowerCAmelCase = "student" , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , )-> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" _UpperCAmelCase = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowerCAmelCase , __lowerCAmelCase ): AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) # purely for convenience _UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(__lowerCAmelCase ).eval() else: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), F"""teacher must be a model or string got type {type(__lowerCAmelCase )}""" _UpperCAmelCase = teacher.config.to_diff_dict() try: _UpperCAmelCase , _UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d} ) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers' ): _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: _UpperCAmelCase , _UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: _UpperCAmelCase = teacher_e if d is None: _UpperCAmelCase = teacher_d if hasattr(teacher.config , 'num_encoder_layers' ): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d} ) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowerCAmelCase ) # Copy weights _UpperCAmelCase = teacher.config_class(**__lowerCAmelCase ) _UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. _UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=__lowerCAmelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save _UpperCAmelCase , _UpperCAmelCase = list(range(__lowerCAmelCase ) ), list(range(__lowerCAmelCase ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(__lowerCAmelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) if d_layers_to_copy is None: _UpperCAmelCase = pick_layers_to_copy(__lowerCAmelCase , __lowerCAmelCase ) try: if hasattr( __lowerCAmelCase , 'prophetnet' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowerCAmelCase ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowerCAmelCase ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowerCAmelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowerCAmelCase ) copy_layers(teacher.decoder.block , student.decoder.block , __lowerCAmelCase ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) _UpperCAmelCase = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowerCAmelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
39
0
from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image lowercase__ =['text', 'image', 'audio'] def __UpperCamelCase ( lowerCAmelCase__ : Any ): __a : Optional[Any] = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((5_1_2, 5_1_2) ) ) elif input_type == "audio": inputs.append(torch.ones(3_0_0_0 ) ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): inputs.append(create_inputs(__lowerCAmelCase ) ) else: raise ValueError(f"Invalid type requested: {input_type}" ) return inputs def __UpperCamelCase ( lowerCAmelCase__ : List[Any] ): __a : Dict = [] for output in outputs: if isinstance(__lowerCAmelCase , (str, AgentText) ): output_types.append('''text''' ) elif isinstance(__lowerCAmelCase , (Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(__lowerCAmelCase , (torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(f"Invalid output: {output}" ) return output_types @is_tool_test class UpperCamelCase__ : def lowerCAmelCase (self : Tuple ): self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) __a : Union[str, Any] = self.tool.inputs for _input in inputs: if isinstance(_input , snake_case_ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) __a : Any = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def lowerCAmelCase (self : Any ): __a : List[Any] = create_inputs(self.tool.inputs ) __a : Optional[int] = self.tool(*snake_case_ ) # There is a single output if len(self.tool.outputs ) == 1: __a : int = [outputs] self.assertListEqual(output_types(snake_case_ ) , self.tool.outputs ) def lowerCAmelCase (self : List[Any] ): self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def lowerCAmelCase (self : Union[str, Any] ): __a : List[Any] = create_inputs(self.tool.inputs ) __a : Optional[Any] = self.tool(*snake_case_ ) if not isinstance(snake_case_ , snake_case_ ): __a : Optional[int] = [outputs] self.assertEqual(len(snake_case_ ) , len(self.tool.outputs ) ) for output, output_type in zip(snake_case_ , self.tool.outputs ): __a : Optional[Any] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(snake_case_ , snake_case_ ) ) def lowerCAmelCase (self : Union[str, Any] ): __a : Dict = create_inputs(self.tool.inputs ) __a : List[str] = [] for _input, input_type in zip(snake_case_ , self.tool.inputs ): if isinstance(snake_case_ , snake_case_ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error __a : List[Any] = self.tool(*snake_case_ ) if not isinstance(snake_case_ , snake_case_ ): __a : Tuple = [outputs] self.assertEqual(len(snake_case_ ) , len(self.tool.outputs ) )
216
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False )-> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCAmelCase = '' else: _UpperCAmelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCAmelCase = in_proj_bias[: config.hidden_size] _UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCAmelCase = in_proj_bias[-config.hidden_size :] def __A ( __lowerCAmelCase )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> int: """simple docstring""" _UpperCAmelCase = dct.pop(__lowerCAmelCase ) _UpperCAmelCase = val def __A ( )-> str: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True )-> List[str]: """simple docstring""" _UpperCAmelCase = ViTConfig() # patch_size if model_name[-1] == "8": _UpperCAmelCase = 8 # set labels if required if not base_model: _UpperCAmelCase = 1_000 _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _UpperCAmelCase = 384 _UpperCAmelCase = 1_536 _UpperCAmelCase = 12 _UpperCAmelCase = 6 # load original model from torch hub _UpperCAmelCase = torch.hub.load('facebookresearch/dino:main' , __lowerCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys _UpperCAmelCase = original_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) _UpperCAmelCase = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if base_model: _UpperCAmelCase = ViTModel(__lowerCAmelCase , add_pooling_layer=__lowerCAmelCase ).eval() else: _UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor _UpperCAmelCase = ViTImageProcessor() _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) _UpperCAmelCase = encoding['pixel_values'] _UpperCAmelCase = model(__lowerCAmelCase ) if base_model: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert torch.allclose(__lowerCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _UpperCAmelCase = original_model(__lowerCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1E-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) _a = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
39
0
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class __magic_name__ : SCREAMING_SNAKE_CASE = 4_2 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None _lowerCAmelCase : Dict = namedtuple("CoinsDistribResult", "moves excess") def UpperCamelCase_( _snake_case : Dict ): """simple docstring""" if root is None: return 0 # Validation def count_nodes(_snake_case : int ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(_snake_case : Tuple ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(__lowerCAmelCase ) != count_coins(__lowerCAmelCase ): raise ValueError('The nodes number should be same as the number of coins' ) # Main calculation def get_distrib(_snake_case : List[str] ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) __a , __a =get_distrib(node.left ) __a , __a =get_distrib(node.right ) __a =1 - left_distrib_excess __a =1 - right_distrib_excess __a =( left_distrib_moves + right_distrib_moves + abs(__lowerCAmelCase ) + abs(__lowerCAmelCase ) ) __a =node.data - coins_to_left - coins_to_right return CoinsDistribResult(__lowerCAmelCase , __lowerCAmelCase ) return get_distrib(__lowerCAmelCase )[0] if __name__ == "__main__": import doctest doctest.testmod()
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
from __future__ import annotations from typing import Any def _UpperCamelCase ( snake_case__ ) -> int: if not postfix_notation: return 0 __UpperCAmelCase : Union[str, Any] = {"+", "-", "*", "/"} __UpperCAmelCase : Optional[int] = [] for token in postfix_notation: if token in operations: __UpperCAmelCase , __UpperCAmelCase : str = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(__lowerCAmelCase ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
157
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=3 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=10 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[1, 1, 2, 1] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase="relu" , UpperCAmelCase=3 , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = embeddings_size _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = scope _UpperCAmelCase = len(UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = TFResNetModel(config=UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFResNetForImageClassification(UpperCAmelCase ) _UpperCAmelCase = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __lowerCamelCase ( snake_case__ , snake_case__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase ( self ): """simple docstring""" pass def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _UpperCAmelCase = model_class(UpperCAmelCase ) _UpperCAmelCase = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase = layer_type _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) @slow def UpperCamelCase ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TFResNetModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __A ( )-> Optional[Any]: """simple docstring""" _UpperCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __lowerCamelCase ( unittest.TestCase): """simple docstring""" @cached_property def UpperCamelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=UpperCAmelCase , return_tensors='tf' ) # forward pass _UpperCAmelCase = model(**UpperCAmelCase ) # verify the logits _UpperCAmelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _UpperCAmelCase = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCAmelCase , atol=1e-4 ) )
39
0
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) class lowerCAmelCase_ ( snake_case__ ): __lowerCamelCase : Dict = "AutoTokenizer" __lowerCamelCase : int = ["tokenizer"] __lowerCamelCase : Dict = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ) -> int: super().__init__(_lowerCAmelCase ) _lowerCAmelCase = speaker_embeddings @classmethod def _snake_case ( cls , _lowerCAmelCase , _lowerCAmelCase="speaker_embeddings_path.json" , **_lowerCAmelCase ) -> Any: if speaker_embeddings_dict_path is not None: _lowerCAmelCase = get_file_from_repo( _lowerCAmelCase , _lowerCAmelCase , subfolder=kwargs.pop("subfolder" , _lowerCAmelCase ) , cache_dir=kwargs.pop("cache_dir" , _lowerCAmelCase ) , force_download=kwargs.pop("force_download" , _lowerCAmelCase ) , proxies=kwargs.pop("proxies" , _lowerCAmelCase ) , resume_download=kwargs.pop("resume_download" , _lowerCAmelCase ) , local_files_only=kwargs.pop("local_files_only" , _lowerCAmelCase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowerCAmelCase ) , revision=kwargs.pop("revision" , _lowerCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowerCAmelCase , _lowerCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _lowerCAmelCase = None else: with open(_lowerCAmelCase ) as speaker_embeddings_json: _lowerCAmelCase = json.load(_lowerCAmelCase ) else: _lowerCAmelCase = None _lowerCAmelCase = AutoTokenizer.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) return cls(tokenizer=_lowerCAmelCase , speaker_embeddings=_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase="speaker_embeddings_path.json" , _lowerCAmelCase="speaker_embeddings" , _lowerCAmelCase = False , **_lowerCAmelCase , ) -> Any: if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowerCAmelCase , _lowerCAmelCase , "v2" ) , exist_ok=_lowerCAmelCase ) _lowerCAmelCase = {} _lowerCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _lowerCAmelCase = self._load_voice_preset(_lowerCAmelCase ) _lowerCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowerCAmelCase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowerCAmelCase , ) _lowerCAmelCase = os.path.join(_lowerCAmelCase , f'''{prompt_key}_{key}.npy''' ) _lowerCAmelCase = tmp_dict with open(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , "w" ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) super().save_pretrained(_lowerCAmelCase , _lowerCAmelCase , **_lowerCAmelCase ) def _snake_case ( self , _lowerCAmelCase = None , **_lowerCAmelCase ) -> str: _lowerCAmelCase = self.speaker_embeddings[voice_preset] _lowerCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _lowerCAmelCase = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowerCAmelCase ) , cache_dir=kwargs.pop("cache_dir" , _lowerCAmelCase ) , force_download=kwargs.pop("force_download" , _lowerCAmelCase ) , proxies=kwargs.pop("proxies" , _lowerCAmelCase ) , resume_download=kwargs.pop("resume_download" , _lowerCAmelCase ) , local_files_only=kwargs.pop("local_files_only" , _lowerCAmelCase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowerCAmelCase ) , revision=kwargs.pop("revision" , _lowerCAmelCase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _lowerCAmelCase = np.load(_lowerCAmelCase ) return voice_preset_dict def _snake_case ( self , _lowerCAmelCase = None ) -> int: for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase="pt" , _lowerCAmelCase=256 , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=False , **_lowerCAmelCase , ) -> str: if voice_preset is not None and not isinstance(_lowerCAmelCase , _lowerCAmelCase ): if ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _lowerCAmelCase = self._load_voice_preset(_lowerCAmelCase ) else: if isinstance(_lowerCAmelCase , _lowerCAmelCase ) and not voice_preset.endswith(".npz" ): _lowerCAmelCase = voice_preset + ".npz" _lowerCAmelCase = np.load(_lowerCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowerCAmelCase , **_lowerCAmelCase ) _lowerCAmelCase = BatchFeature(data=_lowerCAmelCase , tensor_type=_lowerCAmelCase ) _lowerCAmelCase = self.tokenizer( _lowerCAmelCase , return_tensors=_lowerCAmelCase , padding="max_length" , max_length=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , **_lowerCAmelCase , ) if voice_preset is not None: _lowerCAmelCase = voice_preset return encoded_text
158
def __A ( __lowerCAmelCase )-> list: """simple docstring""" if len(__lowerCAmelCase ) < 2: return collection def circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase = False if low == high: return swapped _UpperCAmelCase = low _UpperCAmelCase = high while left < right: if collection[left] > collection[right]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right], collection[left], ) _UpperCAmelCase = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: _UpperCAmelCase , _UpperCAmelCase = ( collection[right + 1], collection[left], ) _UpperCAmelCase = True _UpperCAmelCase = low + int((high - low) / 2 ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , mid + 1 , __lowerCAmelCase ) return swapped or left_swap or right_swap _UpperCAmelCase = True while is_not_sorted is True: _UpperCAmelCase = circle_sort_util(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) - 1 ) return collection if __name__ == "__main__": _a = input('''Enter numbers separated by a comma:\n''').strip() _a = [int(item) for item in user_input.split(''',''')] print(circle_sort(unsorted))
39
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence lowerCAmelCase__ : List[Any] = gray_code_sequence_string(__lowerCAmelCase ) # # convert them to integers for i in range(len(__lowerCAmelCase ) ): lowerCAmelCase__ : Union[str, Any] = int(sequence[i] , 2 ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCAmelCase__ : Any = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCAmelCase__ : Union[str, Any] = gray_code_sequence_string(bit_count - 1 ) lowerCAmelCase__ : Dict = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCAmelCase__ : Optional[int] = '0' + smaller_sequence[i] sequence.append(__lowerCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCAmelCase__ : Optional[Any] = '1' + smaller_sequence[i] sequence.append(__lowerCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
212
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = ["image_processor", "tokenizer"] UpperCamelCase__ = "Pix2StructImageProcessor" UpperCamelCase__ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = False super().__init__(UpperCAmelCase , UpperCAmelCase ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = 2048 , UpperCAmelCase = 0 , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = True , UpperCAmelCase = None , **UpperCAmelCase , ): """simple docstring""" if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , **UpperCAmelCase ) else: # add pixel_values and bbox _UpperCAmelCase = self.image_processor( UpperCAmelCase , return_tensors=UpperCAmelCase , max_patches=UpperCAmelCase , header_text=UpperCAmelCase , **UpperCAmelCase ) if text is not None and not self.image_processor.is_vqa: _UpperCAmelCase = self.tokenizer( text=UpperCAmelCase , add_special_tokens=UpperCAmelCase , padding=UpperCAmelCase , truncation=UpperCAmelCase , max_length=UpperCAmelCase , stride=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_overflowing_tokens=UpperCAmelCase , return_special_tokens_mask=UpperCAmelCase , return_offsets_mapping=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , return_length=UpperCAmelCase , verbose=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase , ) if "attention_mask" in text_encoding: _UpperCAmelCase = text_encoding.pop('attention_mask' ) if "input_ids" in text_encoding: _UpperCAmelCase = text_encoding.pop('input_ids' ) else: _UpperCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase ) return encoding_image_processor def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , *UpperCAmelCase , **UpperCAmelCase ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase , **UpperCAmelCase ) @property def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = self.tokenizer.model_input_names _UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def UpperCAmelCase ( a_ , a_ , a_ ) -> List[str]: """simple docstring""" __A = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") __A = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(__lowerCAmelCase ): os.makedirs(__lowerCAmelCase ) __A = model.state_dict() def to_tf_var_name(a_ ): for patt, repl in iter(__lowerCAmelCase ): __A = name.replace(__lowerCAmelCase , __lowerCAmelCase ) return F'''bert/{name}''' def create_tf_var(a_ , a_ , a_ ): __A = tf.dtypes.as_dtype(tensor.dtype ) __A = tf.get_variable(dtype=__lowerCAmelCase , shape=tensor.shape , name=__lowerCAmelCase , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(__lowerCAmelCase ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __A = to_tf_var_name(__lowerCAmelCase ) __A = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): __A = torch_tensor.T __A = create_tf_var(tensor=__lowerCAmelCase , name=__lowerCAmelCase , session=__lowerCAmelCase ) tf.keras.backend.set_value(__lowerCAmelCase , __lowerCAmelCase ) __A = session.run(__lowerCAmelCase ) print(F'''Successfully created {tf_name}: {np.allclose(__lowerCAmelCase , __lowerCAmelCase )}''' ) __A = tf.train.Saver(tf.trainable_variables() ) saver.save(__lowerCAmelCase , os.path.join(__lowerCAmelCase , model_name.replace("-" , "_" ) + ".ckpt" ) ) def UpperCAmelCase ( a_=None ) -> Any: """simple docstring""" __A = argparse.ArgumentParser() parser.add_argument("--model_name" , type=__lowerCAmelCase , required=__lowerCAmelCase , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=__lowerCAmelCase , default=__lowerCAmelCase , required=__lowerCAmelCase , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=__lowerCAmelCase , required=__lowerCAmelCase , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=__lowerCAmelCase , required=__lowerCAmelCase , help="Directory in which to save tensorflow model" ) __A = parser.parse_args(__lowerCAmelCase ) __A = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=__lowerCAmelCase , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
15
class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase = "" , UpperCAmelCase = False ): """simple docstring""" _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = 0 for q, w in zip(self.prefix , UpperCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" for word in words: self.insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=UpperCAmelCase , is_leaf=UpperCAmelCase ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.nodes.get(word[0] , UpperCAmelCase ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( UpperCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def UpperCamelCase ( self , UpperCAmelCase = 0 ): """simple docstring""" if self.prefix != "": print('-' * height , self.prefix , ' (leaf)' if self.is_leaf else '' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __A ( )-> bool: """simple docstring""" _UpperCAmelCase = 'banana bananas bandana band apple all beast'.split() _UpperCAmelCase = RadixNode() root.insert_many(__lowerCAmelCase ) assert all(root.find(__lowerCAmelCase ) for word in words ) assert not root.find('bandanas' ) assert not root.find('apps' ) root.delete('all' ) assert not root.find('all' ) root.delete('banana' ) assert not root.find('banana' ) assert root.find('bananas' ) return True def __A ( )-> None: """simple docstring""" assert test_trie() def __A ( )-> None: """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = 'banana bananas bandanas bandana band apple all beast'.split() root.insert_many(__lowerCAmelCase ) print('Words:' , __lowerCAmelCase ) print('Tree:' ) root.print_tree() if __name__ == "__main__": main()
39
0
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets __lowerCamelCase : Optional[Any] = '''\ @inproceedings{pillutla-etal:mauve:neurips2021, title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers}, author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid}, booktitle = {NeurIPS}, year = {2021} } ''' __lowerCamelCase : Any = '''\ MAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure. MAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences. For details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021). This metrics is a wrapper around the official implementation of MAUVE: https://github.com/krishnap25/mauve ''' __lowerCamelCase : Optional[Any] = ''' Calculates MAUVE scores between two lists of generated text and reference text. Args: predictions: list of generated text to score. Each predictions should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. Optional Args: num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1 kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9 kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5 kmeans_max_iter: maximum number of k-means iterations. Default 500 featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\']. device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU max_text_length: maximum number of tokens to consider. Default 1024 divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25 mauve_scaling_factor: "c" from the paper. Default 5. verbose: If True (default), print running time updates seed: random seed to initialize k-means cluster assignments. Returns: mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer, frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer, divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve, p_hist: a discrete distribution, which is a quantized version of the text distribution p_text, q_hist: same as above, but with q_text. Examples: >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest >>> import datasets >>> mauve = datasets.load_metric(\'mauve\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP >>> print(out.mauve) # doctest: +SKIP 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): def __a ( self : Optional[Any] ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/krishnap25/mauve""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/krishnap25/mauve"""] , reference_urls=[ """https://arxiv.org/abs/2102.01454""", """https://github.com/krishnap25/mauve""", ] , ) def __a ( self : List[Any] , _lowercase : str , _lowercase : Optional[int] , _lowercase : Union[str, Any]=None , _lowercase : Optional[int]=None , _lowercase : List[Any]=None , _lowercase : Any=None , _lowercase : str="auto" , _lowercase : Optional[int]=-1 , _lowercase : Dict=0.9 , _lowercase : Optional[int]=5 , _lowercase : str=5_00 , _lowercase : Any="gpt2-large" , _lowercase : int=-1 , _lowercase : List[str]=10_24 , _lowercase : int=25 , _lowercase : Optional[Any]=5 , _lowercase : List[Any]=True , _lowercase : Optional[int]=25 , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = compute_mauve( p_text=_lowercase , q_text=_lowercase , p_features=_lowercase , q_features=_lowercase , p_tokens=_lowercase , q_tokens=_lowercase , num_buckets=_lowercase , pca_max_data=_lowercase , kmeans_explained_var=_lowercase , kmeans_num_redo=_lowercase , kmeans_max_iter=_lowercase , featurize_model_name=_lowercase , device_id=_lowercase , max_text_length=_lowercase , divergence_curve_discretization_size=_lowercase , mauve_scaling_factor=_lowercase , verbose=_lowercase , seed=_lowercase , ) return out
219
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _a = 2 class __lowerCamelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments UpperCAmelCase="<s>" , UpperCAmelCase="<pad>" , UpperCAmelCase="</s>" , UpperCAmelCase="<unk>" , UpperCAmelCase=None , ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = bos, unk, pad, eos _UpperCAmelCase = [] _UpperCAmelCase = [] _UpperCAmelCase = {} _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = self.add_symbol(UpperCAmelCase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(UpperCAmelCase ) _UpperCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , UpperCAmelCase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , UpperCAmelCase ): """simple docstring""" return sym in self.indices @classmethod def UpperCamelCase ( cls , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = cls() d.add_from_file(UpperCAmelCase ) return d def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase=1 , UpperCAmelCase=False ): """simple docstring""" if word in self.indices and not overwrite: _UpperCAmelCase = self.indices[word] _UpperCAmelCase = self.count[idx] + n return idx else: _UpperCAmelCase = len(self.symbols ) _UpperCAmelCase = idx self.symbols.append(UpperCAmelCase ) self.count.append(UpperCAmelCase ) return idx def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return 0 def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(UpperCAmelCase ) ) return _UpperCAmelCase = f.readlines() _UpperCAmelCase = self._load_meta(UpperCAmelCase ) for line in lines[indices_start_line:]: try: _UpperCAmelCase , _UpperCAmelCase = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": _UpperCAmelCase = True _UpperCAmelCase , _UpperCAmelCase = line.rsplit(' ' , 1 ) else: _UpperCAmelCase = False _UpperCAmelCase = int(UpperCAmelCase ) _UpperCAmelCase = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(UpperCAmelCase ) ) self.add_symbol(UpperCAmelCase , n=UpperCAmelCase , overwrite=UpperCAmelCase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" _UpperCAmelCase = dict((re.sub(R'@@$' , '' , __lowerCAmelCase ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , __lowerCAmelCase ), v) for k, v in d.items() ) _UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] _UpperCAmelCase = d[k] # restore return da def __A ( __lowerCAmelCase , __lowerCAmelCase )-> str: """simple docstring""" if not os.path.exists(__lowerCAmelCase ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'checkpoint.pt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location='cpu' ) _UpperCAmelCase = chkpt['cfg']['model'] # dicts _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'dict.txt' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) _UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) _UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) _UpperCAmelCase = len(__lowerCAmelCase ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['vocab_file'] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # merges_file (bpecodes) _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'bpecodes' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) _UpperCAmelCase = os.path.join(__lowerCAmelCase , VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowerCAmelCase , __lowerCAmelCase ) # model config _UpperCAmelCase = os.path.join(__lowerCAmelCase , 'config.json' ) _UpperCAmelCase = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1E-12, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # tokenizer config _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 1_024, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(__lowerCAmelCase , ensure_ascii=__lowerCAmelCase , indent=__lowerCAmelCase ) ) # model _UpperCAmelCase = chkpt['model'] # remove unneeded keys _UpperCAmelCase = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) else: _UpperCAmelCase = model_state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase = BioGptConfig.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase = BioGptForCausalLM(__lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(__lowerCAmelCase ) # save _UpperCAmelCase = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) print('Conversion is done!' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _a = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
39
0