code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class lowerCAmelCase : def __init__( self :Optional[int] , _lowercase :Optional[int] , _lowercase :Tuple=13 , _lowercase :Optional[Any]=32 , _lowercase :Optional[Any]=2 , _lowercase :Tuple=3 , _lowercase :Optional[Any]=16 , _lowercase :Union[str, Any]=[1, 2, 1] , _lowercase :Any=[2, 2, 4] , _lowercase :List[Any]=2 , _lowercase :Any=2.0 , _lowercase :List[str]=True , _lowercase :Dict=0.0 , _lowercase :List[str]=0.0 , _lowercase :Optional[Any]=0.1 , _lowercase :Dict="gelu" , _lowercase :Dict=False , _lowercase :Optional[int]=True , _lowercase :str=0.02 , _lowercase :Any=1e-5 , _lowercase :str=True , _lowercase :Union[str, Any]=None , _lowercase :int=True , _lowercase :Tuple=10 , _lowercase :Union[str, Any]=8 , _lowercase :str=["stage1", "stage2", "stage3"] , _lowercase :List[Any]=[1, 2, 3] , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = embed_dim lowercase__ = depths lowercase__ = num_heads lowercase__ = window_size lowercase__ = mlp_ratio lowercase__ = qkv_bias lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = drop_path_rate lowercase__ = hidden_act lowercase__ = use_absolute_embeddings lowercase__ = patch_norm lowercase__ = layer_norm_eps lowercase__ = initializer_range lowercase__ = is_training lowercase__ = scope lowercase__ = use_labels lowercase__ = type_sequence_label_size lowercase__ = encoder_stride lowercase__ = out_features lowercase__ = out_indices def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCAmelCase ( self :Dict , _lowercase :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :str ): '''simple docstring''' lowercase__ = MaskFormerSwinModel(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) lowercase__ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowercase__ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :Optional[int] , _lowercase :List[Any] , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = MaskFormerSwinBackbone(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_lowercase ): lowercase__ = ["stem"] lowercase__ = MaskFormerSwinBackbone(config=_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __lowerCamelCase = {'feature-extraction': MaskFormerSwinModel} if is_torch_available() else {} __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = MaskFormerSwinModelTester(self ) lowercase__ = ConfigTester(self , config_class=_lowercase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( "`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with" " `nn.DataParallel`" ) ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' pass def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self :str ): '''simple docstring''' return def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowercase ) @unittest.skip("Swin does not use inputs_embeds" ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' pass @unittest.skip("Swin does not support feedforward chunking" ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' pass def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowercase , nn.Linear ) ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) @unittest.skip(reason="MaskFormerSwin is only used as backbone and doesn't support output_attentions" ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' pass @unittest.skip(reason="MaskFormerSwin is only used as an internal backbone" ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' pass def UpperCAmelCase ( self :List[Any] , _lowercase :Optional[int] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.eval() with torch.no_grad(): lowercase__ = model(**self._prepare_for_class(_lowercase , _lowercase ) ) lowercase__ = outputs.hidden_states lowercase__ = getattr( self.model_tester , "expected_num_hidden_layers" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowercase ) , _lowercase ) # Swin has a different seq_length lowercase__ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowercase__ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowercase__ = True self.check_hidden_states_output(_lowercase , _lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ = True self.check_hidden_states_output(_lowercase , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = 3 lowercase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowercase__ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowercase__ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowercase__ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowercase__ = True self.check_hidden_states_output(_lowercase , _lowercase , _lowercase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ = True self.check_hidden_states_output(_lowercase , _lowercase , _lowercase , (padded_height, padded_width) ) @unittest.skip(reason="MaskFormerSwin doesn't have pretrained checkpoints" ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' pass @unittest.skip(reason="This will be fixed once MaskFormerSwin is replaced by native Swin" ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' pass @unittest.skip(reason="This will be fixed once MaskFormerSwin is replaced by native Swin" ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' pass def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_lowercase :List[Any] ): lowercase__ = 0 return t def check_equivalence(_lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Optional[Any] , _lowercase :str={} ): with torch.no_grad(): lowercase__ = model(**_lowercase , return_dict=_lowercase , **_lowercase ) lowercase__ = model(**_lowercase , return_dict=_lowercase , **_lowercase ).to_tuple() def recursive_check(_lowercase :List[str] , _lowercase :str ): if isinstance(_lowercase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowercase , _lowercase ): recursive_check(_lowercase , _lowercase ) elif isinstance(_lowercase , _lowercase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_lowercase , _lowercase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_lowercase ) , set_nan_tensor_to_zero(_lowercase ) , atol=1e-5 ) , msg=( "Tuple and dict output are not equal. Difference:" f''' {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:''' f''' {torch.isnan(_lowercase ).any()} and `inf`: {torch.isinf(_lowercase )}. Dict has''' f''' `nan`: {torch.isnan(_lowercase ).any()} and `inf`: {torch.isinf(_lowercase )}.''' ) , ) recursive_check(_lowercase , _lowercase ) for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = self._prepare_for_class(_lowercase , _lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase ) check_equivalence(_lowercase , _lowercase , _lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) check_equivalence(_lowercase , _lowercase , _lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase ) check_equivalence(_lowercase , _lowercase , _lowercase , {"output_hidden_states": True} ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) check_equivalence(_lowercase , _lowercase , _lowercase , {"output_hidden_states": True} ) @require_torch class lowerCAmelCase ( unittest.TestCase , lowercase_ ): __lowerCamelCase = (MaskFormerSwinBackbone,) if is_torch_available() else () __lowerCamelCase = MaskFormerSwinConfig def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = MaskFormerSwinModelTester(self ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = inputs_dict["pixel_values"].shape[0] for backbone_class in self.all_model_classes: lowercase__ = backbone_class(_lowercase ) backbone.to(_lowercase ) backbone.eval() lowercase__ = backbone(**_lowercase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _lowercase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True lowercase__ = backbone(**_lowercase , output_hidden_states=_lowercase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) lowercase__ , lowercase__ , lowercase__ = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: lowercase__ = backbone(**_lowercase , output_attentions=_lowercase ) self.assertIsNotNone(outputs.attentions )
655
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
1
from ....utils import logging _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :List[Any] , _lowercase :Union[str, Any]=None , _lowercase :Optional[Any]=20_48 ): '''simple docstring''' lowercase__ = config.__dict__ lowercase__ = modal_hidden_size if num_labels: lowercase__ = num_labels
655
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
1
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = KandinskyVaaPipeline __lowerCamelCase = [ 'image_embeds', 'negative_image_embeds', ] __lowerCamelCase = ['image_embeds', 'negative_image_embeds'] __lowerCamelCase = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] __lowerCamelCase = False @property def UpperCAmelCase ( self :int ): '''simple docstring''' return 32 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return 32 @property def UpperCAmelCase ( self :str ): '''simple docstring''' return self.time_input_dim @property def UpperCAmelCase ( self :int ): '''simple docstring''' return self.time_input_dim * 4 @property def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' return 1_00 @property def UpperCAmelCase ( self :Any ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } lowercase__ = UNetaDConditionModel(**_lowercase ) return model @property def UpperCAmelCase ( self :str ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.dummy_unet lowercase__ = self.dummy_movq lowercase__ = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule="linear" , beta_start=0.00085 , beta_end=0.012 , clip_sample=_lowercase , set_alpha_to_one=_lowercase , steps_offset=1 , prediction_type="epsilon" , thresholding=_lowercase , ) lowercase__ = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def UpperCAmelCase ( self :Any , _lowercase :str , _lowercase :Optional[int]=0 ): '''simple docstring''' lowercase__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowercase ) ).to(_lowercase ) lowercase__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowercase ) if str(_lowercase ).startswith("mps" ): lowercase__ = torch.manual_seed(_lowercase ) else: lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = "cpu" lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = pipe(**self.get_dummy_inputs(_lowercase ) ) lowercase__ = output.images lowercase__ = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] lowercase__ = image[0, -3:, -3:, -1] lowercase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase__ = np.array( [0.6237976, 1.0, 0.36441332, 1.0, 0.70639634, 0.29877186, 0.85652125, 0.5216843, 0.54454046] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) lowercase__ = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(_lowercase ) lowercase__ = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) lowercase__ = pipeline.to(_lowercase ) pipeline.set_progress_bar_config(disable=_lowercase ) lowercase__ = "red cat, 4k photo" lowercase__ = torch.Generator(device="cuda" ).manual_seed(0 ) lowercase__ , lowercase__ = pipe_prior( _lowercase , generator=_lowercase , num_inference_steps=5 , negative_prompt="" , ).to_tuple() lowercase__ = torch.Generator(device="cuda" ).manual_seed(0 ) lowercase__ = pipeline( image_embeds=_lowercase , negative_image_embeds=_lowercase , generator=_lowercase , num_inference_steps=1_00 , output_type="np" , ) lowercase__ = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(_lowercase , _lowercase )
655
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["prompt"] lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] if "image" in inputs: lowercase__ = inputs["image"] else: lowercase__ = None if "mask_image" in inputs: lowercase__ = inputs["mask_image"] else: lowercase__ = None if "original_image" in inputs: lowercase__ = inputs["original_image"] else: lowercase__ = None lowercase__ , lowercase__ = pipe.encode_prompt(_lowercase ) # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_lowercase , _lowercase , _lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowercase , _lowercase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 )
655
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """weiweishi/roc-bert-base-zh""": """https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'roc_bert' def __init__( self :List[str] , _lowercase :int=3_05_22 , _lowercase :Optional[Any]=7_68 , _lowercase :Union[str, Any]=12 , _lowercase :List[str]=12 , _lowercase :List[str]=30_72 , _lowercase :Optional[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :List[str]=0.1 , _lowercase :str=5_12 , _lowercase :List[str]=2 , _lowercase :int=0.02 , _lowercase :Optional[Any]=1e-12 , _lowercase :Tuple=True , _lowercase :Tuple=0 , _lowercase :Dict="absolute" , _lowercase :Optional[int]=None , _lowercase :Tuple=True , _lowercase :List[Any]=True , _lowercase :Dict=7_68 , _lowercase :Any=9_10 , _lowercase :List[str]=5_12 , _lowercase :Optional[Any]=2_48_58 , _lowercase :Union[str, Any]=True , **_lowercase :Dict , ): '''simple docstring''' lowercase__ = vocab_size lowercase__ = max_position_embeddings lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = initializer_range lowercase__ = type_vocab_size lowercase__ = layer_norm_eps lowercase__ = use_cache lowercase__ = enable_pronunciation lowercase__ = enable_shape lowercase__ = pronunciation_embed_dim lowercase__ = pronunciation_vocab_size lowercase__ = shape_embed_dim lowercase__ = shape_vocab_size lowercase__ = concat_input lowercase__ = position_embedding_type lowercase__ = classifier_dropout super().__init__(pad_token_id=_lowercase , **_lowercase )
655
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) lowercase__ = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase__ = model(_lowercase )["last_hidden_state"] lowercase__ = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
655
1
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
1
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 # setable values __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None @classmethod def UpperCAmelCase ( cls :Union[str, Any] , _lowercase :CommonSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray ): '''simple docstring''' return cls(common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase ) @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCamelCase = 42 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return True @register_to_config def __init__( self :str , _lowercase :int = 10_00 , _lowercase :float = 0.0001 , _lowercase :float = 0.02 , _lowercase :str = "linear" , _lowercase :Optional[jnp.ndarray] = None , _lowercase :str = "fixed_small" , _lowercase :bool = True , _lowercase :str = "epsilon" , _lowercase :jnp.dtype = jnp.floataa , ): '''simple docstring''' lowercase__ = dtype def UpperCAmelCase ( self :str , _lowercase :Optional[CommonSchedulerState] = None ): '''simple docstring''' if common is None: lowercase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowercase__ = jnp.array(1.0 , dtype=self.dtype ) lowercase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase ( self :List[str] , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :Tuple = () ): '''simple docstring''' lowercase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowercase__ = (jnp.arange(0 , _lowercase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Tuple , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :List[str]=None , _lowercase :Tuple=None ): '''simple docstring''' lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowercase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowercase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowercase__ = jnp.clip(_lowercase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowercase__ = jnp.log(jnp.clip(_lowercase , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowercase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowercase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowercase__ = variance lowercase__ = state.common.betas[t] lowercase__ = (predicted_variance + 1) / 2 lowercase__ = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase ( self :Optional[int] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :int , _lowercase :jnp.ndarray , _lowercase :Optional[jax.random.KeyArray] = None , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = timestep if key is None: lowercase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ = jnp.split(_lowercase , sample.shape[1] , axis=1 ) else: lowercase__ = None # 1. compute alphas, betas lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowercase__ = 1 - alpha_prod_t lowercase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowercase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowercase__ = model_output elif self.config.prediction_type == "v_prediction": lowercase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowercase__ = jnp.clip(_lowercase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowercase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowercase__ = jax.random.split(_lowercase , num=1 ) lowercase__ = jax.random.normal(_lowercase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowercase , _lowercase , predicted_variance=_lowercase ) ** 0.5) * noise lowercase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowercase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowercase , state=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return add_noise_common(state.common , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return get_velocity_common(state.common , _lowercase , _lowercase , _lowercase ) def __len__( self :List[str] ): '''simple docstring''' return self.config.num_train_timesteps
655
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
1
import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) _snake_case = logging.getLogger(__name__) _snake_case = """Hello world! cécé herlolip""" _snake_case = namedtuple( """BertAbsConfig""", [ """temp_dir""", """large""", """use_bert_emb""", """finetune_bert""", """encoder""", """share_emb""", """max_pos""", """enc_layers""", """enc_hidden_size""", """enc_heads""", """enc_ff_size""", """enc_dropout""", """dec_layers""", """dec_hidden_size""", """dec_heads""", """dec_ff_size""", """dec_dropout""", ], ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = BertAbsConfig( temp_dir="." , finetune_bert=__magic_name__ , large=__magic_name__ , share_emb=__magic_name__ , use_bert_emb=__magic_name__ , encoder="bert" , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) lowercase__ = torch.load(__magic_name__ , lambda __magic_name__ , __magic_name__ : storage ) lowercase__ = AbsSummarizer(__magic_name__ , torch.device("cpu" ) , __magic_name__ ) original.eval() lowercase__ = BertAbsSummarizer(__magic_name__ , torch.device("cpu" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("convert the model" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("Make sure that the models' outputs are identical" ) lowercase__ = BertTokenizer.from_pretrained("bert-base-uncased" ) # prepare the model inputs lowercase__ = tokenizer.encode("This is sample éàalj'-." ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(__magic_name__ )) ) lowercase__ = torch.tensor(__magic_name__ ).unsqueeze(0 ) lowercase__ = tokenizer.encode("This is sample 3 éàalj'-." ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(__magic_name__ )) ) lowercase__ = torch.tensor(__magic_name__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass lowercase__ = encoder_input_ids lowercase__ = decoder_input_ids lowercase__ = lowercase__ = None lowercase__ = None lowercase__ = lowercase__ = None lowercase__ = lowercase__ = None lowercase__ = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical lowercase__ = original(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ )[0] lowercase__ = original.generator(__magic_name__ ) lowercase__ = new_model( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ )[0] lowercase__ = new_model.generator(__magic_name__ ) lowercase__ = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(__magic_name__ ) ) lowercase__ = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(__magic_name__ ) ) lowercase__ = torch.allclose(__magic_name__ , __magic_name__ , atol=1e-3 ) if are_identical: logging.info("all weights are equal up to 1e-3" ) else: raise ValueError("the weights are different. The new model is likely different from the original one." ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("saving the model's state dictionary" ) torch.save( new_model.state_dict() , "./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--bertabs_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""", ) _snake_case = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
655
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = SMALL_MODEL_IDENTIFIER lowercase__ = "pt" lowercase__ = "tf" def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :int ): '''simple docstring''' lowercase__ = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = "mock_framework" # Framework provided - return whatever the user provides lowercase__ = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model )
655
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { """configuration_ctrl""": ["""CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CTRLConfig"""], """tokenization_ctrl""": ["""CTRLTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """CTRL_PRETRAINED_MODEL_ARCHIVE_LIST""", """CTRLForSequenceClassification""", """CTRLLMHeadModel""", """CTRLModel""", """CTRLPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFCTRLForSequenceClassification""", """TFCTRLLMHeadModel""", """TFCTRLModel""", """TFCTRLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git_vision_model' def __init__( self :Dict , _lowercase :Dict=7_68 , _lowercase :Dict=30_72 , _lowercase :Tuple=12 , _lowercase :List[str]=12 , _lowercase :Tuple=3 , _lowercase :Dict=2_24 , _lowercase :Tuple=16 , _lowercase :Optional[int]="quick_gelu" , _lowercase :Union[str, Any]=1e-5 , _lowercase :Tuple=0.0 , _lowercase :Tuple=0.02 , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = hidden_size lowercase__ = intermediate_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_channels lowercase__ = patch_size lowercase__ = image_size lowercase__ = initializer_range lowercase__ = attention_dropout lowercase__ = layer_norm_eps lowercase__ = hidden_act @classmethod def UpperCAmelCase ( cls :List[str] , _lowercase :Union[str, os.PathLike] , **_lowercase :Optional[int] ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git' def __init__( self :Union[str, Any] , _lowercase :Dict=None , _lowercase :List[str]=3_05_22 , _lowercase :Tuple=7_68 , _lowercase :Any=6 , _lowercase :Dict=12 , _lowercase :Any=30_72 , _lowercase :List[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :Optional[int]=0.1 , _lowercase :Optional[Any]=10_24 , _lowercase :Any=0.02 , _lowercase :int=1e-12 , _lowercase :List[Any]=0 , _lowercase :int="absolute" , _lowercase :List[str]=True , _lowercase :Any=False , _lowercase :int=1_01 , _lowercase :str=1_02 , _lowercase :Dict=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , pad_token_id=_lowercase , **_lowercase ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowercase__ = GitVisionConfig(**_lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = tie_word_embeddings lowercase__ = num_image_with_embedding lowercase__ = bos_token_id lowercase__ = eos_token_id def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
655
1
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def _A ( __magic_name__ ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def _A ( __magic_name__ ): class lowerCAmelCase : def __init__( self :Optional[int] , _lowercase :Dict ): '''simple docstring''' lowercase__ = metric_id class lowerCAmelCase : __lowerCamelCase = [MetricMock(lowercase_ ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): if "tmp_path" in args: lowercase__ = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(__magic_name__ , match="https://huggingface.co/docs/evaluate" ): func(*__magic_name__ )
655
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
1
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _snake_case = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val def _A ( __magic_name__ ): lowercase__ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowercase__ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) lowercase__ = value else: lowercase__ = value return new_state_dict def _A ( __magic_name__ ): lowercase__ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowercase__ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowercase__ = in_proj_weight_cross_attn[:256, :] lowercase__ = in_proj_bias_cross_attn[:256] lowercase__ = in_proj_weight_cross_attn[256:512, :] lowercase__ = in_proj_bias_cross_attn[256:512] lowercase__ = in_proj_weight_cross_attn[-256:, :] lowercase__ = in_proj_bias_cross_attn[-256:] def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = image.size lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 800 if "detection" in checkpoint_url else 1000 lowercase__ = target_max_size / current_max_size lowercase__ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _A ( __magic_name__ ): lowercase__ = F.to_tensor(__magic_name__ ) lowercase__ = F.normalize(__magic_name__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): logger.info("Converting model..." ) # load original state dict lowercase__ = torch.hub.load_state_dict_from_url(__magic_name__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowercase__ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val # create HuggingFace model and load state dict lowercase__ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: lowercase__ = 15 lowercase__ = 2 lowercase__ = {0: "table", 1: "table rotated"} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 125 lowercase__ = 6 lowercase__ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) lowercase__ = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion lowercase__ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" lowercase__ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=__magic_name__ ) lowercase__ = Image.open(__magic_name__ ).convert("RGB" ) lowercase__ = normalize(resize(__magic_name__ , __magic_name__ ) ).unsqueeze(0 ) lowercase__ = model(__magic_name__ ) if "detection" in checkpoint_url: lowercase__ = (1, 15, 3) lowercase__ = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) lowercase__ = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: lowercase__ = (1, 125, 7) lowercase__ = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) lowercase__ = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __magic_name__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) lowercase__ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _snake_case = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
655
1
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = None if token is not None: lowercase__ = {"Accept": "application/vnd.github+json", "Authorization": f'''Bearer {token}'''} lowercase__ = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' lowercase__ = requests.get(__magic_name__ , headers=__magic_name__ ).json() lowercase__ = {} try: job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) lowercase__ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(__magic_name__ ): lowercase__ = requests.get(url + f'''&page={i + 2}''' , headers=__magic_name__ ).json() job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) return job_links except Exception: print(f'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = None if token is not None: lowercase__ = {"Accept": "application/vnd.github+json", "Authorization": f'''Bearer {token}'''} lowercase__ = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100''' lowercase__ = requests.get(__magic_name__ , headers=__magic_name__ ).json() lowercase__ = {} try: artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) lowercase__ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(__magic_name__ ): lowercase__ = requests.get(url + f'''&page={i + 2}''' , headers=__magic_name__ ).json() artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) return artifacts except Exception: print(f'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = None if token is not None: lowercase__ = {"Accept": "application/vnd.github+json", "Authorization": f'''Bearer {token}'''} lowercase__ = requests.get(__magic_name__ , headers=__magic_name__ , allow_redirects=__magic_name__ ) lowercase__ = result.headers["Location"] lowercase__ = requests.get(__magic_name__ , allow_redirects=__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , f'''{artifact_name}.zip''' ) with open(__magic_name__ , "wb" ) as fp: fp.write(response.content ) def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = [] lowercase__ = [] lowercase__ = None with zipfile.ZipFile(__magic_name__ ) as z: for filename in z.namelist(): if not os.path.isdir(__magic_name__ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(__magic_name__ ) as f: for line in f: lowercase__ = line.decode("UTF-8" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs lowercase__ = line[: line.index(": " )] lowercase__ = line[line.index(": " ) + len(": " ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("FAILED " ): # `test` is the test method that failed lowercase__ = line[len("FAILED " ) :] failed_tests.append(__magic_name__ ) elif filename == "job_name.txt": lowercase__ = line if len(__magic_name__ ) != len(__magic_name__ ): raise ValueError( f'''`errors` and `failed_tests` should have the same number of elements. Got {len(__magic_name__ )} for `errors` ''' f'''and {len(__magic_name__ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some''' " problem." ) lowercase__ = None if job_name and job_links: lowercase__ = job_links.get(__magic_name__ , __magic_name__ ) # A list with elements of the form (line of error, error, failed test) lowercase__ = [x + [y] + [job_link] for x, y in zip(__magic_name__ , __magic_name__ )] return result def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = [] lowercase__ = [os.path.join(__magic_name__ , __magic_name__ ) for p in os.listdir(__magic_name__ ) if p.endswith(".zip" )] for p in paths: errors.extend(get_errors_from_single_artifact(__magic_name__ , job_links=__magic_name__ ) ) return errors def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = Counter() counter.update([x[1] for x in logs] ) lowercase__ = counter.most_common() lowercase__ = {} for error, count in counts: if error_filter is None or error not in error_filter: lowercase__ = {"count": count, "failed_tests": [(x[2], x[0]) for x in logs if x[1] == error]} lowercase__ = dict(sorted(r.items() , key=lambda __magic_name__ : item[1]["count"] , reverse=__magic_name__ ) ) return r def _A ( __magic_name__ ): lowercase__ = test.split("::" )[0] if test.startswith("tests/models/" ): lowercase__ = test.split("/" )[2] else: lowercase__ = None return test def _A ( __magic_name__ , __magic_name__=None ): lowercase__ = [(x[0], x[1], get_model(x[2] )) for x in logs] lowercase__ = [x for x in logs if x[2] is not None] lowercase__ = {x[2] for x in logs} lowercase__ = {} for test in tests: lowercase__ = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) lowercase__ = counter.most_common() lowercase__ = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} lowercase__ = sum(error_counts.values() ) if n_errors > 0: lowercase__ = {"count": n_errors, "errors": error_counts} lowercase__ = dict(sorted(r.items() , key=lambda __magic_name__ : item[1]["count"] , reverse=__magic_name__ ) ) return r def _A ( __magic_name__ ): lowercase__ = "| no. | error | status |" lowercase__ = "|-:|:-|:-|" lowercase__ = [header, sep] for error in reduced_by_error: lowercase__ = reduced_by_error[error]["count"] lowercase__ = f'''| {count} | {error[:100]} | |''' lines.append(__magic_name__ ) return "\n".join(__magic_name__ ) def _A ( __magic_name__ ): lowercase__ = "| model | no. of errors | major error | count |" lowercase__ = "|-:|-:|-:|-:|" lowercase__ = [header, sep] for model in reduced_by_model: lowercase__ = reduced_by_model[model]["count"] lowercase__ , lowercase__ = list(reduced_by_model[model]["errors"].items() )[0] lowercase__ = f'''| {model} | {count} | {error[:60]} | {_count} |''' lines.append(__magic_name__ ) return "\n".join(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") _snake_case = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) _snake_case = get_job_links(args.workflow_run_id, token=args.token) _snake_case = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: _snake_case = k.find(""" / """) _snake_case = k[index + len(""" / """) :] _snake_case = v with open(os.path.join(args.output_dir, """job_links.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) _snake_case = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) _snake_case = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error _snake_case = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors _snake_case = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, """errors.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) _snake_case = reduce_by_error(errors) _snake_case = reduce_by_model(errors) _snake_case = make_github_table(reduced_by_error) _snake_case = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, """reduced_by_error.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa) with open(os.path.join(args.output_dir, """reduced_by_model.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa)
655
from typing import TYPE_CHECKING from ...utils import _LazyModule _snake_case = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
1
from __future__ import annotations def _A ( __magic_name__ ): # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(__magic_name__ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(__magic_name__ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
655
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = """ Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _A ( __magic_name__ , __magic_name__ , __magic_name__=8 ): lowercase__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :UNetaDConditionModel , _lowercase :DDPMScheduler , _lowercase :VQModel , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) lowercase__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :int , _lowercase :str ): '''simple docstring''' if latents is None: lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase__ = latents.to(_lowercase ) lowercase__ = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self :int , _lowercase :int=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) lowercase__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase__ = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase__ , lowercase__ = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. lowercase__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__( self :int , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 1_00 , _lowercase :float = 4.0 , _lowercase :int = 1 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = self._execution_device lowercase__ = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) lowercase__ = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: lowercase__ = image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) lowercase__ = self.scheduler.timesteps lowercase__ = self.unet.config.in_channels lowercase__ , lowercase__ = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) # create initial latent lowercase__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowercase , _lowercase , _lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance lowercase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase__ = {"image_embeds": image_embeds} lowercase__ = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) lowercase__ , lowercase__ = noise_pred.chunk(2 ) lowercase__ , lowercase__ = variance_pred.chunk(2 ) lowercase__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase__ = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing lowercase__ = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase__ = image * 0.5 + 0.5 lowercase__ = image.clamp(0 , 1 ) lowercase__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
655
1
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES _snake_case = logging.get_logger(__name__) _snake_case = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) _snake_case = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) _snake_case = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) _snake_case = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) _snake_case = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) _snake_case = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) _snake_case = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) _snake_case = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) _snake_case = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) _snake_case = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) _snake_case = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) _snake_case = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) _snake_case = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) _snake_case = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) _snake_case = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_MAPPING _snake_case = auto_class_update(FlaxAutoModel) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_PRETRAINING_MAPPING _snake_case = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING _snake_case = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_MASKED_LM_MAPPING _snake_case = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _snake_case = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _snake_case = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING _snake_case = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING _snake_case = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING _snake_case = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING _snake_case = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _snake_case = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING _snake_case = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class lowerCAmelCase ( _BaseAutoModelClass ): __lowerCamelCase = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING _snake_case = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
655
import inspect import unittest class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :int ): '''simple docstring''' try: import diffusers # noqa: F401 except ImportError: assert False def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' import diffusers from diffusers.dependency_versions_table import deps lowercase__ = inspect.getmembers(_lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowercase__ = "k-diffusion" elif backend == "invisible_watermark": lowercase__ = "invisible-watermark" assert backend in deps, f'''{backend} is not in the deps table!'''
655
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _snake_case = logging.get_logger(__name__) _snake_case = { """Helsinki-NLP/opus-mt-en-de""": """https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json""", # See all Marian models at https://huggingface.co/models?filter=marian } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'marian' __lowerCamelCase = ['past_key_values'] __lowerCamelCase = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self :int , _lowercase :List[Any]=5_81_01 , _lowercase :int=None , _lowercase :Tuple=10_24 , _lowercase :int=12 , _lowercase :List[Any]=40_96 , _lowercase :Optional[int]=16 , _lowercase :Optional[Any]=12 , _lowercase :Any=40_96 , _lowercase :Union[str, Any]=16 , _lowercase :str=0.0 , _lowercase :Dict=0.0 , _lowercase :List[Any]=True , _lowercase :List[Any]=True , _lowercase :Union[str, Any]="gelu" , _lowercase :Tuple=10_24 , _lowercase :int=0.1 , _lowercase :Optional[int]=0.0 , _lowercase :Union[str, Any]=0.0 , _lowercase :Union[str, Any]=0.02 , _lowercase :Optional[int]=5_81_00 , _lowercase :Dict=False , _lowercase :int=5_81_00 , _lowercase :Union[str, Any]=0 , _lowercase :int=0 , _lowercase :str=True , **_lowercase :List[Any] , ): '''simple docstring''' lowercase__ = vocab_size lowercase__ = decoder_vocab_size or vocab_size lowercase__ = max_position_embeddings lowercase__ = d_model lowercase__ = encoder_ffn_dim lowercase__ = encoder_layers lowercase__ = encoder_attention_heads lowercase__ = decoder_ffn_dim lowercase__ = decoder_layers lowercase__ = decoder_attention_heads lowercase__ = dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = activation_function lowercase__ = init_std lowercase__ = encoder_layerdrop lowercase__ = decoder_layerdrop lowercase__ = use_cache lowercase__ = encoder_layers lowercase__ = scale_embedding # scale factor will be sqrt(d_model) if True lowercase__ = share_encoder_decoder_embeddings super().__init__( pad_token_id=_lowercase , eos_token_id=_lowercase , is_encoder_decoder=_lowercase , decoder_start_token_id=_lowercase , forced_eos_token_id=_lowercase , **_lowercase , ) class lowerCAmelCase ( lowercase_ ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: lowercase__ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: lowercase__ = {0: "batch"} lowercase__ = {0: "batch", 1: "past_decoder_sequence + sequence"} else: lowercase__ = {0: "batch", 1: "decoder_sequence"} lowercase__ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_lowercase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase__ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: lowercase__ , lowercase__ = self.num_layers for i in range(_lowercase ): lowercase__ = {0: "batch", 2: "past_sequence + sequence"} lowercase__ = {0: "batch", 2: "past_sequence + sequence"} else: lowercase__ = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def UpperCAmelCase ( self :Any ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: lowercase__ = super().outputs else: lowercase__ = super(_lowercase , self ).outputs if self.use_past: lowercase__ , lowercase__ = self.num_layers for i in range(_lowercase ): lowercase__ = {0: "batch", 2: "past_sequence + sequence"} lowercase__ = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def UpperCAmelCase ( self :List[Any] , _lowercase :PreTrainedTokenizer , _lowercase :int = -1 , _lowercase :int = -1 , _lowercase :bool = False , _lowercase :Optional[TensorType] = None , ): '''simple docstring''' lowercase__ = self._generate_dummy_inputs_for_encoder_and_decoder( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) # Generate decoder inputs lowercase__ = seq_length if not self.use_past else 1 lowercase__ = self._generate_dummy_inputs_for_encoder_and_decoder( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) lowercase__ = {f'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} lowercase__ = dict(**_lowercase , **_lowercase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowercase__ , lowercase__ = common_inputs["input_ids"].shape lowercase__ = common_inputs["decoder_input_ids"].shape[1] lowercase__ , lowercase__ = self.num_attention_heads lowercase__ = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ = decoder_seq_length + 3 lowercase__ = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase__ = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(_lowercase , _lowercase )] , dim=1 ) lowercase__ = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase__ , lowercase__ = self.num_layers lowercase__ = min(_lowercase , _lowercase ) lowercase__ = max(_lowercase , _lowercase ) - min_num_layers lowercase__ = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(_lowercase ): common_inputs["past_key_values"].append( ( torch.zeros(_lowercase ), torch.zeros(_lowercase ), torch.zeros(_lowercase ), torch.zeros(_lowercase ), ) ) # TODO: test this. lowercase__ = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(_lowercase , _lowercase ): common_inputs["past_key_values"].append((torch.zeros(_lowercase ), torch.zeros(_lowercase )) ) return common_inputs def UpperCAmelCase ( self :str , _lowercase :PreTrainedTokenizer , _lowercase :int = -1 , _lowercase :int = -1 , _lowercase :bool = False , _lowercase :Optional[TensorType] = None , ): '''simple docstring''' lowercase__ = self._generate_dummy_inputs_for_encoder_and_decoder( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowercase__ , lowercase__ = common_inputs["input_ids"].shape # Not using the same length for past_key_values lowercase__ = seqlen + 2 lowercase__ , lowercase__ = self.num_layers lowercase__ , lowercase__ = self.num_attention_heads lowercase__ = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ = common_inputs["attention_mask"].dtype lowercase__ = torch.cat( [common_inputs["attention_mask"], torch.ones(_lowercase , _lowercase , dtype=_lowercase )] , dim=1 ) lowercase__ = [ (torch.zeros(_lowercase ), torch.zeros(_lowercase )) for _ in range(_lowercase ) ] return common_inputs def UpperCAmelCase ( self :Tuple , _lowercase :PreTrainedTokenizer , _lowercase :int = -1 , _lowercase :int = -1 , _lowercase :bool = False , _lowercase :Optional[TensorType] = None , ): '''simple docstring''' lowercase__ = compute_effective_axis_dimension( _lowercase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase__ = tokenizer.num_special_tokens_to_add(_lowercase ) lowercase__ = compute_effective_axis_dimension( _lowercase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_lowercase ) # Generate dummy inputs according to compute batch and sequence lowercase__ = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase__ = dict(tokenizer(_lowercase , return_tensors=_lowercase ) ) return common_inputs def UpperCAmelCase ( self :str , _lowercase :PreTrainedTokenizer , _lowercase :int = -1 , _lowercase :int = -1 , _lowercase :bool = False , _lowercase :Optional[TensorType] = None , ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: lowercase__ = self._generate_dummy_inputs_for_default_and_seqaseq_lm( _lowercase , batch_size=_lowercase , seq_length=_lowercase , is_pair=_lowercase , framework=_lowercase ) else: lowercase__ = self._generate_dummy_inputs_for_causal_lm( _lowercase , batch_size=_lowercase , seq_length=_lowercase , is_pair=_lowercase , framework=_lowercase ) return common_inputs def UpperCAmelCase ( self :Tuple , _lowercase :Tuple , _lowercase :Optional[int] , _lowercase :Dict , _lowercase :Tuple ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: lowercase__ = super()._flatten_past_key_values_(_lowercase , _lowercase , _lowercase , _lowercase ) else: lowercase__ = super(_lowercase , self )._flatten_past_key_values_( _lowercase , _lowercase , _lowercase , _lowercase ) @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return 1e-4
655
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 # setable values __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None @classmethod def UpperCAmelCase ( cls :Union[str, Any] , _lowercase :CommonSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray ): '''simple docstring''' return cls(common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase ) @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCamelCase = 42 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return True @register_to_config def __init__( self :str , _lowercase :int = 10_00 , _lowercase :float = 0.0001 , _lowercase :float = 0.02 , _lowercase :str = "linear" , _lowercase :Optional[jnp.ndarray] = None , _lowercase :str = "fixed_small" , _lowercase :bool = True , _lowercase :str = "epsilon" , _lowercase :jnp.dtype = jnp.floataa , ): '''simple docstring''' lowercase__ = dtype def UpperCAmelCase ( self :str , _lowercase :Optional[CommonSchedulerState] = None ): '''simple docstring''' if common is None: lowercase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowercase__ = jnp.array(1.0 , dtype=self.dtype ) lowercase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase ( self :List[str] , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :Tuple = () ): '''simple docstring''' lowercase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowercase__ = (jnp.arange(0 , _lowercase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Tuple , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :List[str]=None , _lowercase :Tuple=None ): '''simple docstring''' lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowercase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowercase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowercase__ = jnp.clip(_lowercase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowercase__ = jnp.log(jnp.clip(_lowercase , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowercase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowercase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowercase__ = variance lowercase__ = state.common.betas[t] lowercase__ = (predicted_variance + 1) / 2 lowercase__ = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase ( self :Optional[int] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :int , _lowercase :jnp.ndarray , _lowercase :Optional[jax.random.KeyArray] = None , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = timestep if key is None: lowercase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ = jnp.split(_lowercase , sample.shape[1] , axis=1 ) else: lowercase__ = None # 1. compute alphas, betas lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowercase__ = 1 - alpha_prod_t lowercase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowercase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowercase__ = model_output elif self.config.prediction_type == "v_prediction": lowercase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowercase__ = jnp.clip(_lowercase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowercase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowercase__ = jax.random.split(_lowercase , num=1 ) lowercase__ = jax.random.normal(_lowercase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowercase , _lowercase , predicted_variance=_lowercase ) ** 0.5) * noise lowercase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowercase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowercase , state=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return add_noise_common(state.common , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return get_velocity_common(state.common , _lowercase , _lowercase , _lowercase ) def __len__( self :List[str] ): '''simple docstring''' return self.config.num_train_timesteps
655
1
from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ['input_features', 'attention_mask'] def __init__( self :Optional[int] , _lowercase :Optional[int]=80 , _lowercase :int=1_60_00 , _lowercase :Any=80 , _lowercase :Union[str, Any]=0.0 , _lowercase :int=True , _lowercase :Tuple=True , _lowercase :int=True , **_lowercase :int , ): '''simple docstring''' super().__init__(feature_size=_lowercase , sampling_rate=_lowercase , padding_value=_lowercase , **_lowercase ) lowercase__ = num_mel_bins lowercase__ = do_ceptral_normalize lowercase__ = normalize_means lowercase__ = normalize_vars lowercase__ = True def UpperCAmelCase ( self :Tuple , _lowercase :np.ndarray , ): '''simple docstring''' lowercase__ = waveform * (2**15) # Kaldi compliance: 16-bit signed integers lowercase__ = torch.from_numpy(_lowercase ).unsqueeze(0 ) lowercase__ = ta_kaldi.fbank(_lowercase , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def UpperCAmelCase ( _lowercase :np.ndarray , _lowercase :int , _lowercase :Optional[bool] = True , _lowercase :Optional[bool] = True , _lowercase :float = 0.0 , ): '''simple docstring''' if normalize_means: lowercase__ = x[:input_length].mean(axis=0 ) lowercase__ = np.subtract(_lowercase , _lowercase ) if normalize_vars: lowercase__ = x[:input_length].std(axis=0 ) lowercase__ = np.divide(_lowercase , _lowercase ) if input_length < x.shape[0]: lowercase__ = padding_value # make sure array is in float32 lowercase__ = x.astype(np.floataa ) return x def UpperCAmelCase ( self :Optional[Any] , _lowercase :List[np.ndarray] , _lowercase :Optional[np.ndarray] = None ): '''simple docstring''' lowercase__ = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(_lowercase , _lowercase , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(_lowercase , _lowercase ) ] def __call__( self :int , _lowercase :Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , _lowercase :Union[bool, str, PaddingStrategy] = False , _lowercase :Optional[int] = None , _lowercase :bool = False , _lowercase :Optional[int] = None , _lowercase :Optional[Union[str, TensorType]] = None , _lowercase :Optional[int] = None , _lowercase :Optional[bool] = None , **_lowercase :Optional[Any] , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowercase__ = isinstance(_lowercase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowercase__ = is_batched_numpy or ( isinstance(_lowercase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowercase__ = [np.asarray(_lowercase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_lowercase , np.ndarray ): lowercase__ = np.asarray(_lowercase , dtype=np.floataa ) elif isinstance(_lowercase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowercase__ = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowercase__ = [raw_speech] # extract fbank features lowercase__ = [self._extract_fbank_features(_lowercase ) for waveform in raw_speech] # convert into correct format for padding lowercase__ = BatchFeature({"input_features": features} ) lowercase__ = self.pad( _lowercase , padding=_lowercase , max_length=_lowercase , truncation=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) # make sure list is in array format lowercase__ = padded_inputs.get("input_features" ) if isinstance(input_features[0] , _lowercase ): lowercase__ = [np.asarray(_lowercase , dtype=np.floataa ) for feature in input_features] lowercase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowercase__ = [np.asarray(_lowercase , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: lowercase__ = ( np.array(_lowercase , dtype=np.intaa ) if self._get_padding_strategies(_lowercase , max_length=_lowercase ) is not PaddingStrategy.DO_NOT_PAD else None ) lowercase__ = self.normalize( padded_inputs["input_features"] , attention_mask=_lowercase ) if return_tensors is not None: lowercase__ = padded_inputs.convert_to_tensors(_lowercase ) return padded_inputs
655
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _snake_case = logging.get_logger(__name__) _snake_case = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _snake_case = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCAmelCase : __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Model type selected in the list: ' + ', '.join(lowercase_ )} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) __lowerCamelCase = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) __lowerCamelCase = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) __lowerCamelCase = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) __lowerCamelCase = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'train' __lowerCamelCase = 'dev' class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 def __init__( self :Optional[Any] , _lowercase :SquadDataTrainingArguments , _lowercase :PreTrainedTokenizer , _lowercase :Optional[int] = None , _lowercase :Union[str, Split] = Split.train , _lowercase :Optional[bool] = False , _lowercase :Optional[str] = None , _lowercase :Optional[str] = "pt" , ): '''simple docstring''' lowercase__ = args lowercase__ = is_language_sensitive lowercase__ = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_lowercase , _lowercase ): try: lowercase__ = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) lowercase__ = mode # Load data features from cache or dataset file lowercase__ = "v2" if args.version_2_with_negative else "v1" lowercase__ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase__ = cached_features_file + ".lock" with FileLock(_lowercase ): if os.path.exists(_lowercase ) and not args.overwrite_cache: lowercase__ = time.time() lowercase__ = torch.load(_lowercase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowercase__ = self.old_features["features"] lowercase__ = self.old_features.get("dataset" , _lowercase ) lowercase__ = self.old_features.get("examples" , _lowercase ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: lowercase__ = self.processor.get_dev_examples(args.data_dir ) else: lowercase__ = self.processor.get_train_examples(args.data_dir ) lowercase__ , lowercase__ = squad_convert_examples_to_features( examples=self.examples , tokenizer=_lowercase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_lowercase , ) lowercase__ = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _lowercase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self :Dict ): '''simple docstring''' return len(self.features ) def __getitem__( self :Any , _lowercase :Any ): '''simple docstring''' lowercase__ = self.features[i] lowercase__ = torch.tensor(feature.input_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.attention_mask , dtype=torch.long ) lowercase__ = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.cls_index , dtype=torch.long ) lowercase__ = torch.tensor(feature.p_mask , dtype=torch.float ) lowercase__ = torch.tensor(feature.is_impossible , dtype=torch.float ) lowercase__ = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowercase__ = torch.tensor(feature.start_position , dtype=torch.long ) lowercase__ = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
655
1
import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def _A ( __magic_name__ ): lowercase__ = [ "decoder.version", "decoder.output_projection.weight", "_float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(__magic_name__ , __magic_name__ ) def _A ( __magic_name__ ): lowercase__ , lowercase__ = emb.weight.shape lowercase__ = nn.Linear(__magic_name__ , __magic_name__ , bias=__magic_name__ ) lowercase__ = emb.weight.data return lin_layer def _A ( __magic_name__ ): lowercase__ = torch.load(__magic_name__ , map_location="cpu" ) lowercase__ = Namespace(**checkpoint["cfg"]["model"] ) lowercase__ = checkpoint["model"] remove_ignore_keys_(__magic_name__ ) lowercase__ = state_dict["decoder.embed_tokens.weight"].shape[0] lowercase__ = {key.replace("decoder" , "model" ): val for key, val in state_dict.items()} lowercase__ = XGLMConfig( vocab_size=__magic_name__ , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="gelu" , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) lowercase__ = XGLMForCausalLM(__magic_name__ ) lowercase__ = model.load_state_dict(__magic_name__ , strict=__magic_name__ ) print(__magic_name__ ) lowercase__ = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _snake_case = parser.parse_args() _snake_case = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
655
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = """▁""" _snake_case = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} _snake_case = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } _snake_case = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } _snake_case = { """ernie-m-base""": 514, """ernie-m-large""": 514, } _snake_case = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ["input_ids"] __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = RESOURCE_FILES_NAMES def __init__( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=False , _lowercase :Dict="utf8" , _lowercase :Optional[Any]="[UNK]" , _lowercase :Optional[int]="[SEP]" , _lowercase :List[str]="[PAD]" , _lowercase :Dict="[CLS]" , _lowercase :Optional[Any]="[MASK]" , _lowercase :Optional[Dict[str, Any]] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , vocab_file=_lowercase , encoding=_lowercase , sp_model_kwargs=self.sp_model_kwargs , **_lowercase , ) lowercase__ = do_lower_case lowercase__ = sentencepiece_model_ckpt lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase__ = self.load_vocab(filepath=_lowercase ) else: lowercase__ = {self.sp_model.id_to_piece(_lowercase ): id for id in range(self.sp_model.get_piece_size() )} lowercase__ = {v: k for k, v in self.vocab.items()} def UpperCAmelCase ( self :Any , _lowercase :Dict ): '''simple docstring''' if text is None: return None lowercase__ = self.tokenize(_lowercase ) lowercase__ , lowercase__ = "", [] for i, ch in enumerate(_lowercase ): if ch in self.SP_CHAR_MAPPING: lowercase__ = self.SP_CHAR_MAPPING.get(_lowercase ) else: lowercase__ = unicodedata.normalize("NFKC" , _lowercase ) if self.is_whitespace(_lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(_lowercase ) ) lowercase__ , lowercase__ , lowercase__ = normalized_text, [], 0 if self.do_lower_case: lowercase__ = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase__ = token[1:] lowercase__ = text[offset:].index(_lowercase ) + offset lowercase__ = start + len(_lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase__ = end return token_mapping @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return len(self.vocab ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self :Any ): '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self :Optional[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCAmelCase ( self :List[str] , _lowercase :Optional[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(_lowercase , _lowercase ) for c in text) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any]=False , _lowercase :Optional[int]=64 , _lowercase :Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get("enable_sampling" ) is True: lowercase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: lowercase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: lowercase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: lowercase__ = self.sp_model.EncodeAsPieces(_lowercase ) else: lowercase__ = self.sp_model.SampleEncodeAsPieces(_lowercase , _lowercase , _lowercase ) lowercase__ = [] for pi, piece in enumerate(_lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(_lowercase ) and pi != 0: new_pieces.append(_lowercase ) continue else: continue lowercase__ = 0 for i, chunk in enumerate(_lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(_lowercase ) or self.is_punct(_lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(_lowercase ) lowercase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i if len(_lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCAmelCase ( self :Tuple , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = self.convert_ids_to_tokens(_lowercase ) lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' return self.vocab.get(_lowercase , self.vocab.get(self.unk_token ) ) def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' return self.reverse_vocab.get(_lowercase , self.unk_token ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :Tuple=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCAmelCase ( self :Dict , _lowercase :int , _lowercase :Union[str, Any]=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Dict=None , _lowercase :Optional[Any]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(_lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(_lowercase ) + 1) + [1] * (len(_lowercase ) + 3) def UpperCAmelCase ( self :str , _lowercase :Optional[int] ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCAmelCase ( self :int , _lowercase :Dict ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCAmelCase ( self :List[str] , _lowercase :List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(_lowercase ) == 1: lowercase__ = unicodedata.category(_lowercase ) if cat == "Zs": return True return False def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = {} with io.open(_lowercase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(_lowercase ): lowercase__ = line.rstrip("\n" ) lowercase__ = int(_lowercase ) return token_to_idx def UpperCAmelCase ( self :List[str] , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' lowercase__ = 0 if os.path.isdir(_lowercase ): lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowercase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(_lowercase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowercase__ = token_index writer.write(token + "\n" ) index += 1 lowercase__ = os.path.join(_lowercase , "sentencepiece.bpe.model" ) with open(_lowercase , "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(_lowercase ) return (vocab_file,)
655
1
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _snake_case = logging.getLogger(__name__) @dataclass(frozen=lowercase_ ) class lowerCAmelCase : __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None @dataclass(frozen=lowercase_ ) class lowerCAmelCase : __lowerCamelCase = 42 __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if is_torch_available(): import torch from torch.utils.data import Dataset class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 def __init__( self :Dict , _lowercase :str , _lowercase :PreTrainedTokenizer , _lowercase :str , _lowercase :Optional[int] = None , _lowercase :str=False , _lowercase :bool = False , ): '''simple docstring''' lowercase__ = hans_processors[task]() lowercase__ = os.path.join( _lowercase , "cached_{}_{}_{}_{}".format( "dev" if evaluate else "train" , tokenizer.__class__.__name__ , str(_lowercase ) , _lowercase , ) , ) lowercase__ = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowercase__ , lowercase__ = label_list[2], label_list[1] lowercase__ = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase__ = cached_features_file + ".lock" with FileLock(_lowercase ): if os.path.exists(_lowercase ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) lowercase__ = torch.load(_lowercase ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) lowercase__ = ( processor.get_dev_examples(_lowercase ) if evaluate else processor.get_train_examples(_lowercase ) ) logger.info("Training examples: %s" , len(_lowercase ) ) lowercase__ = hans_convert_examples_to_features(_lowercase , _lowercase , _lowercase , _lowercase ) logger.info("Saving features into cached file %s" , _lowercase ) torch.save(self.features , _lowercase ) def __len__( self :Union[str, Any] ): '''simple docstring''' return len(self.features ) def __getitem__( self :int , _lowercase :Optional[Any] ): '''simple docstring''' return self.features[i] def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return self.label_list if is_tf_available(): import tensorflow as tf class lowerCAmelCase : __lowerCamelCase = 42 def __init__( self :Any , _lowercase :str , _lowercase :PreTrainedTokenizer , _lowercase :str , _lowercase :Optional[int] = 1_28 , _lowercase :Tuple=False , _lowercase :bool = False , ): '''simple docstring''' lowercase__ = hans_processors[task]() lowercase__ = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowercase__ , lowercase__ = label_list[2], label_list[1] lowercase__ = label_list lowercase__ = processor.get_dev_examples(_lowercase ) if evaluate else processor.get_train_examples(_lowercase ) lowercase__ = hans_convert_examples_to_features(_lowercase , _lowercase , _lowercase , _lowercase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc="convert examples to features" ): if ex_index % 1_00_00 == 0: logger.info("Writing example %d of %d" % (ex_index, len(_lowercase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowercase__ = tf.data.Dataset.from_generator( _lowercase , ( { "example_id": tf.intaa, "input_ids": tf.intaa, "attention_mask": tf.intaa, "token_type_ids": tf.intaa, }, tf.intaa, ) , ( { "example_id": tf.TensorShape([] ), "input_ids": tf.TensorShape([None, None] ), "attention_mask": tf.TensorShape([None, None] ), "token_type_ids": tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return self.dataset def __len__( self :int ): '''simple docstring''' return len(self.features ) def __getitem__( self :List[Any] , _lowercase :Union[str, Any] ): '''simple docstring''' return self.features[i] def UpperCAmelCase ( self :str ): '''simple docstring''' return self.label_list class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Dict , _lowercase :Tuple ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(_lowercase , "heuristics_train_set.txt" ) ) , "train" ) def UpperCAmelCase ( self :Optional[int] , _lowercase :int ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(_lowercase , "heuristics_evaluation_set.txt" ) ) , "dev" ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return ["contradiction", "entailment", "neutral"] def UpperCAmelCase ( self :int , _lowercase :Union[str, Any] , _lowercase :str ): '''simple docstring''' lowercase__ = [] for i, line in enumerate(_lowercase ): if i == 0: continue lowercase__ = "%s-%s" % (set_type, line[0]) lowercase__ = line[5] lowercase__ = line[6] lowercase__ = line[7][2:] if line[7].startswith("ex" ) else line[7] lowercase__ = line[0] examples.append(InputExample(guid=_lowercase , text_a=_lowercase , text_b=_lowercase , label=_lowercase , pairID=_lowercase ) ) return examples def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ): lowercase__ = {label: i for i, label in enumerate(__magic_name__ )} lowercase__ = [] for ex_index, example in tqdm.tqdm(enumerate(__magic_name__ ) , desc="convert examples to features" ): if ex_index % 1_0000 == 0: logger.info("Writing example %d" % (ex_index) ) lowercase__ = tokenizer( example.text_a , example.text_b , add_special_tokens=__magic_name__ , max_length=__magic_name__ , padding="max_length" , truncation=__magic_name__ , return_overflowing_tokens=__magic_name__ , ) lowercase__ = label_map[example.label] if example.label in label_map else 0 lowercase__ = int(example.pairID ) features.append(InputFeatures(**__magic_name__ , label=__magic_name__ , pairID=__magic_name__ ) ) for i, example in enumerate(examples[:5] ): logger.info("*** Example ***" ) logger.info(f'''guid: {example}''' ) logger.info(f'''features: {features[i]}''' ) return features _snake_case = { """hans""": 3, } _snake_case = { """hans""": HansProcessor, }
655
def _A ( __magic_name__ ): lowercase__ = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def _A ( __magic_name__ = 100 ): lowercase__ = 1 lowercase__ = 2 for i in range(2 , max_n + 1 ): lowercase__ = pre_numerator lowercase__ = 2 * i // 3 if i % 3 == 0 else 1 lowercase__ = cur_numerator lowercase__ = e_cont * pre_numerator + temp return sum_digits(__magic_name__ ) if __name__ == "__main__": print(F"""{solution() = }""")
655
1
def _A ( __magic_name__ ): if any(not isinstance(__magic_name__ , __magic_name__ ) or x < 0 for x in sequence ): raise TypeError("Sequence must be list of non-negative integers" ) for _ in range(len(__magic_name__ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(__magic_name__ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
655
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'AutoTokenizer' __lowerCamelCase = ['tokenizer'] __lowerCamelCase = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self :Dict , _lowercase :List[str] , _lowercase :List[Any]=None ): '''simple docstring''' super().__init__(_lowercase ) lowercase__ = speaker_embeddings @classmethod def UpperCAmelCase ( cls :Any , _lowercase :int , _lowercase :str="speaker_embeddings_path.json" , **_lowercase :List[str] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase__ = get_file_from_repo( _lowercase , _lowercase , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowercase , _lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase__ = None else: with open(_lowercase ) as speaker_embeddings_json: lowercase__ = json.load(_lowercase ) else: lowercase__ = None lowercase__ = AutoTokenizer.from_pretrained(_lowercase , **_lowercase ) return cls(tokenizer=_lowercase , speaker_embeddings=_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :List[str]="speaker_embeddings_path.json" , _lowercase :Any="speaker_embeddings" , _lowercase :bool = False , **_lowercase :Any , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowercase , _lowercase , "v2" ) , exist_ok=_lowercase ) lowercase__ = {} lowercase__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase__ = self._load_voice_preset(_lowercase ) lowercase__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowercase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowercase , ) lowercase__ = os.path.join(_lowercase , f'''{prompt_key}_{key}.npy''' ) lowercase__ = tmp_dict with open(os.path.join(_lowercase , _lowercase ) , "w" ) as fp: json.dump(_lowercase , _lowercase ) super().save_pretrained(_lowercase , _lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :str = None , **_lowercase :List[Any] ): '''simple docstring''' lowercase__ = self.speaker_embeddings[voice_preset] lowercase__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase__ = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) lowercase__ = np.load(_lowercase ) return voice_preset_dict def UpperCAmelCase ( self :Optional[int] , _lowercase :Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Optional[Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=None , _lowercase :List[str]="pt" , _lowercase :List[Any]=2_56 , _lowercase :List[str]=False , _lowercase :Union[str, Any]=True , _lowercase :Dict=False , **_lowercase :Tuple , ): '''simple docstring''' if voice_preset is not None and not isinstance(_lowercase , _lowercase ): if ( isinstance(_lowercase , _lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase__ = self._load_voice_preset(_lowercase ) else: if isinstance(_lowercase , _lowercase ) and not voice_preset.endswith(".npz" ): lowercase__ = voice_preset + ".npz" lowercase__ = np.load(_lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowercase , **_lowercase ) lowercase__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) lowercase__ = self.tokenizer( _lowercase , return_tensors=_lowercase , padding="max_length" , max_length=_lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , add_special_tokens=_lowercase , **_lowercase , ) if voice_preset is not None: lowercase__ = voice_preset return encoded_text
655
1
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version _snake_case = logging.getLogger(__name__) require_version("""pytorch_lightning>=1.0.4""") _snake_case = { """base""": AutoModel, """sequence-classification""": AutoModelForSequenceClassification, """question-answering""": AutoModelForQuestionAnswering, """pretraining""": AutoModelForPreTraining, """token-classification""": AutoModelForTokenClassification, """language-modeling""": AutoModelWithLMHead, """summarization""": AutoModelForSeqaSeqLM, """translation""": AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization _snake_case = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } _snake_case = sorted(arg_to_scheduler.keys()) _snake_case = """{""" + """, """.join(arg_to_scheduler_choices) + """}""" class lowerCAmelCase ( pl.LightningModule ): def __init__( self :List[Any] , _lowercase :argparse.Namespace , _lowercase :Optional[Any]=None , _lowercase :int="base" , _lowercase :str=None , _lowercase :List[Any]=None , _lowercase :Optional[Any]=None , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(_lowercase ) lowercase__ = 0 lowercase__ = Path(self.hparams.output_dir ) lowercase__ = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: lowercase__ = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({"num_labels": num_labels} if num_labels is not None else {}) , cache_dir=_lowercase , **_lowercase , ) else: lowercase__ = config lowercase__ = ("encoder_layerdrop", "decoder_layerdrop", "dropout", "attention_dropout") for p in extra_model_params: if getattr(self.hparams , _lowercase , _lowercase ): assert hasattr(self.config , _lowercase ), f'''model config doesn\'t have a `{p}` attribute''' setattr(self.config , _lowercase , getattr(self.hparams , _lowercase ) ) if tokenizer is None: lowercase__ = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=_lowercase , ) else: lowercase__ = tokenizer lowercase__ = MODEL_MODES[mode] if model is None: lowercase__ = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool(".ckpt" in self.hparams.model_name_or_path ) , config=self.config , cache_dir=_lowercase , ) else: lowercase__ = model def UpperCAmelCase ( self :Optional[Any] , *_lowercase :Optional[Any] , **_lowercase :Optional[int] ): '''simple docstring''' lowercase__ = self.model_type.from_pretrained(*_lowercase , **_lowercase ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = arg_to_scheduler[self.hparams.lr_scheduler] lowercase__ = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps() ) lowercase__ = {"scheduler": scheduler, "interval": "step", "frequency": 1} return scheduler def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.model lowercase__ = ["bias", "LayerNorm.weight"] lowercase__ = [ { "params": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters "weight_decay": self.hparams.weight_decay, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], "weight_decay": 0.0, }, ] if self.hparams.adafactor: lowercase__ = Adafactor( _lowercase , lr=self.hparams.learning_rate , scale_parameter=_lowercase , relative_step=_lowercase ) else: lowercase__ = AdamW( _lowercase , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon ) lowercase__ = optimizer lowercase__ = self.get_lr_scheduler() return [optimizer], [scheduler] def UpperCAmelCase ( self :Dict , _lowercase :Any , _lowercase :Any ): '''simple docstring''' return self.validation_step(_lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :Dict ): '''simple docstring''' return self.validation_end(_lowercase ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = max(1 , self.hparams.gpus ) # TODO: consider num_tpu_cores lowercase__ = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def UpperCAmelCase ( self :Optional[Any] , _lowercase :Optional[Any] ): '''simple docstring''' if stage == "test": lowercase__ = len(self.test_dataloader().dataset ) else: lowercase__ = self.get_dataloader("train" , self.hparams.train_batch_size , shuffle=_lowercase ) lowercase__ = len(self.train_dataloader().dataset ) def UpperCAmelCase ( self :str , _lowercase :str , _lowercase :int , _lowercase :bool = False ): '''simple docstring''' raise NotImplementedError("You must implement this for your task" ) def UpperCAmelCase ( self :str ): '''simple docstring''' return self.train_loader def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self.get_dataloader("dev" , self.hparams.eval_batch_size , shuffle=_lowercase ) def UpperCAmelCase ( self :str ): '''simple docstring''' return self.get_dataloader("test" , self.hparams.eval_batch_size , shuffle=_lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :Any ): '''simple docstring''' return os.path.join( self.hparams.data_dir , "cached_{}_{}_{}".format( _lowercase , list(filter(_lowercase , self.hparams.model_name_or_path.split("/" ) ) ).pop() , str(self.hparams.max_seq_length ) , ) , ) @pl.utilities.rank_zero_only def UpperCAmelCase ( self :List[str] , _lowercase :Dict[str, Any] ): '''simple docstring''' lowercase__ = self.output_dir.joinpath("best_tfmr" ) lowercase__ = self.step_count self.model.save_pretrained(_lowercase ) self.tokenizer.save_pretrained(_lowercase ) @staticmethod def UpperCAmelCase ( _lowercase :Optional[int] , _lowercase :Any ): '''simple docstring''' parser.add_argument( "--model_name_or_path" , default=_lowercase , type=_lowercase , required=_lowercase , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--config_name" , default="" , type=_lowercase , help="Pretrained config name or path if not the same as model_name" ) parser.add_argument( "--tokenizer_name" , default=_lowercase , type=_lowercase , help="Pretrained tokenizer name or path if not the same as model_name" , ) parser.add_argument( "--cache_dir" , default=str(Path(_lowercase ).parent / "test_run" / "cache" ) , type=_lowercase , help="Where do you want to store the pre-trained models downloaded from huggingface.co" , ) parser.add_argument( "--encoder_layerdrop" , type=_lowercase , help="Encoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--decoder_layerdrop" , type=_lowercase , help="Decoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--dropout" , type=_lowercase , help="Dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--attention_dropout" , type=_lowercase , help="Attention dropout probability (Optional). Goes into model.config" , ) parser.add_argument("--learning_rate" , default=5e-5 , type=_lowercase , help="The initial learning rate for Adam." ) parser.add_argument( "--lr_scheduler" , default="linear" , choices=_lowercase , metavar=_lowercase , type=_lowercase , help="Learning rate scheduler" , ) parser.add_argument("--weight_decay" , default=0.0 , type=_lowercase , help="Weight decay if we apply some." ) parser.add_argument("--adam_epsilon" , default=1e-8 , type=_lowercase , help="Epsilon for Adam optimizer." ) parser.add_argument("--warmup_steps" , default=0 , type=_lowercase , help="Linear warmup over warmup_steps." ) parser.add_argument("--num_workers" , default=4 , type=_lowercase , help="kwarg passed to DataLoader" ) parser.add_argument("--num_train_epochs" , dest="max_epochs" , default=3 , type=_lowercase ) parser.add_argument("--train_batch_size" , default=32 , type=_lowercase ) parser.add_argument("--eval_batch_size" , default=32 , type=_lowercase ) parser.add_argument("--adafactor" , action="store_true" ) class lowerCAmelCase ( pl.Callback ): def UpperCAmelCase ( self :Optional[Any] , _lowercase :str , _lowercase :str ): '''simple docstring''' if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class lowerCAmelCase ( pl.Callback ): def UpperCAmelCase ( self :Dict , _lowercase :Tuple , _lowercase :Optional[int] ): '''simple docstring''' for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(_lowercase ) class lowerCAmelCase ( pl.Callback ): def UpperCAmelCase ( self :List[str] , _lowercase :Tuple , _lowercase :str ): '''simple docstring''' lowercase__ = trainer.lr_schedulers[0]["scheduler"] lowercase__ = {f'''lr_group_{i}''': lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(_lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :pl.Trainer , _lowercase :pl.LightningModule ): '''simple docstring''' rank_zero_info("***** Validation results *****" ) lowercase__ = trainer.callback_metrics # Log results for key in sorted(_lowercase ): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(_lowercase , str(metrics[key] ) ) ) def UpperCAmelCase ( self :Any , _lowercase :pl.Trainer , _lowercase :pl.LightningModule ): '''simple docstring''' rank_zero_info("***** Test results *****" ) lowercase__ = trainer.callback_metrics # Log and save results to file lowercase__ = os.path.join(pl_module.hparams.output_dir , "test_results.txt" ) with open(_lowercase , "w" ) as writer: for key in sorted(_lowercase ): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(_lowercase , str(metrics[key] ) ) ) writer.write("{} = {}\n".format(_lowercase , str(metrics[key] ) ) ) def _A ( __magic_name__ , __magic_name__ ): # To allow all pl args uncomment the following line # parser = pl.Trainer.add_argparse_args(parser) parser.add_argument( "--output_dir" , default=str(Path(__magic_name__ ).parent / "test_run" / "model_checkpoints" ) , type=__magic_name__ , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument( "--fp16" , action="store_true" , help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit" , ) parser.add_argument( "--fp16_opt_level" , type=__magic_name__ , default="O2" , help=( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']." "See details at https://nvidia.github.io/apex/amp.html" ) , ) parser.add_argument("--n_tpu_cores" , dest="tpu_cores" , type=__magic_name__ ) parser.add_argument("--max_grad_norm" , dest="gradient_clip_val" , default=1.0 , type=__magic_name__ , help="Max gradient norm" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_predict" , action="store_true" , help="Whether to run predictions on the test set." ) parser.add_argument( "--gradient_accumulation_steps" , dest="accumulate_grad_batches" , type=__magic_name__ , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--seed" , type=__magic_name__ , default=42 , help="random seed for initialization" ) parser.add_argument( "--data_dir" , default=str(Path(__magic_name__ ).parent / "test_run" / "dummy-train-data" ) , type=__magic_name__ , help="The input data dir. Should contain the training files for the CoNLL-2003 NER task." , ) def _A ( __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=True , __magic_name__=[] , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): pl.seed_everything(args.seed ) # init model lowercase__ = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=__magic_name__ ) # add custom checkpoints if checkpoint_callback is None: lowercase__ = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="checkpoint" , monitor="val_loss" , mode="min" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(__magic_name__ ) if logging_callback is None: lowercase__ = LoggingCallback() lowercase__ = {} if args.fpaa: lowercase__ = 16 if args.gpus > 1: lowercase__ = "auto" lowercase__ = "ddp" lowercase__ = args.accumulate_grad_batches lowercase__ = None lowercase__ = "auto" lowercase__ = pl.Trainer.from_argparse_args( __magic_name__ , weights_summary=__magic_name__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=__magic_name__ , val_check_interval=1 , num_sanity_val_steps=2 , **__magic_name__ , ) if args.do_train: trainer.fit(__magic_name__ ) else: print("RAG modeling tests with new set functions successfuly executed!" ) return trainer
655
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__magic_name__ ): # Forward propagation lowercase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? lowercase__ = (expected / 100) - layer_a # Error delta lowercase__ = layer_1_error * sigmoid_function(__magic_name__ , __magic_name__ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input("""Expected value: """)) _snake_case = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
655
1
import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_lowercase , "width_multiplier" ) ) class lowerCAmelCase : def __init__( self :Any , _lowercase :str , _lowercase :Any=13 , _lowercase :Any=64 , _lowercase :Optional[int]=2 , _lowercase :List[str]=3 , _lowercase :Tuple="swish" , _lowercase :int=3 , _lowercase :str=32 , _lowercase :Tuple=0.1 , _lowercase :int=0.02 , _lowercase :str=True , _lowercase :Tuple=True , _lowercase :List[str]=10 , _lowercase :Optional[int]=None , _lowercase :Dict=0.25 , _lowercase :List[Any]=0.0 , _lowercase :str=0.0 , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = make_divisible(5_12 * width_multiplier , divisor=8 ) lowercase__ = hidden_act lowercase__ = conv_kernel_size lowercase__ = output_stride lowercase__ = classifier_dropout_prob lowercase__ = use_labels lowercase__ = is_training lowercase__ = num_labels lowercase__ = initializer_range lowercase__ = scope lowercase__ = width_multiplier lowercase__ = ffn_dropout lowercase__ = attn_dropout def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.num_labels ) lowercase__ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowercase__ = self.get_config() return config, pixel_values, labels, pixel_labels def UpperCAmelCase ( self :Dict ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :List[Any] , _lowercase :Optional[Any] , _lowercase :Tuple , _lowercase :Any ): '''simple docstring''' lowercase__ = MobileViTVaModel(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCAmelCase ( self :int , _lowercase :Union[str, Any] , _lowercase :str , _lowercase :Optional[Any] , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = MobileViTVaForImageClassification(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self :str , _lowercase :List[str] , _lowercase :List[Any] , _lowercase :str , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = MobileViTVaForSemanticSegmentation(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowercase__ = model(_lowercase , labels=_lowercase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __lowerCamelCase = ( { 'feature-extraction': MobileViTVaModel, 'image-classification': MobileViTVaForImageClassification, 'image-segmentation': MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = MobileViTVaModelTester(self ) lowercase__ = MobileViTVaConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileViTV2 does not use inputs_embeds" ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="MobileViTV2 does not support input and output embeddings" ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' pass @unittest.skip(reason="MobileViTV2 does not output attentions" ) def UpperCAmelCase ( self :int ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason="Got `CUDA error: misaligned address` for tests after this one being run." ) def UpperCAmelCase ( self :Any ): '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' pass def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' def check_hidden_states_output(_lowercase :Any , _lowercase :Optional[int] , _lowercase :Union[str, Any] ): lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.eval() with torch.no_grad(): lowercase__ = model(**self._prepare_for_class(_lowercase , _lowercase ) ) lowercase__ = outputs.hidden_states lowercase__ = 5 self.assertEqual(len(_lowercase ) , _lowercase ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. lowercase__ = 2 for i in range(len(_lowercase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowercase ) @slow def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = MobileViTVaModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def _A ( ): lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCAmelCase ( unittest.TestCase ): @cached_property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained("apple/mobilevitv2-1.0-imagenet1k-256" ) if is_vision_available() else None ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = MobileViTVaForImageClassification.from_pretrained("apple/mobilevitv2-1.0-imagenet1k-256" ).to( _lowercase ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): lowercase__ = model(**_lowercase ) # verify the logits lowercase__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _lowercase ) lowercase__ = torch.tensor([-1.63_36e00, -7.32_04e-02, -5.18_83e-01] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = MobileViTVaForSemanticSegmentation.from_pretrained("shehan97/mobilevitv2-1.0-voc-deeplabv3" ) lowercase__ = model.to(_lowercase ) lowercase__ = MobileViTImageProcessor.from_pretrained("shehan97/mobilevitv2-1.0-voc-deeplabv3" ) lowercase__ = prepare_img() lowercase__ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): lowercase__ = model(**_lowercase ) lowercase__ = outputs.logits # verify the logits lowercase__ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , _lowercase ) lowercase__ = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=_lowercase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , _lowercase , atol=1e-4 ) ) @slow def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = MobileViTVaForSemanticSegmentation.from_pretrained("shehan97/mobilevitv2-1.0-voc-deeplabv3" ) lowercase__ = model.to(_lowercase ) lowercase__ = MobileViTImageProcessor.from_pretrained("shehan97/mobilevitv2-1.0-voc-deeplabv3" ) lowercase__ = prepare_img() lowercase__ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): lowercase__ = model(**_lowercase ) lowercase__ = outputs.logits.detach().cpu() lowercase__ = image_processor.post_process_semantic_segmentation(outputs=_lowercase , target_sizes=[(50, 60)] ) lowercase__ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , _lowercase ) lowercase__ = image_processor.post_process_semantic_segmentation(outputs=_lowercase ) lowercase__ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , _lowercase )
655
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'van' def __init__( self :Optional[Any] , _lowercase :Dict=2_24 , _lowercase :Union[str, Any]=3 , _lowercase :List[Any]=[7, 3, 3, 3] , _lowercase :Any=[4, 2, 2, 2] , _lowercase :Union[str, Any]=[64, 1_28, 3_20, 5_12] , _lowercase :List[Any]=[3, 3, 12, 3] , _lowercase :Dict=[8, 8, 4, 4] , _lowercase :int="gelu" , _lowercase :List[Any]=0.02 , _lowercase :List[Any]=1e-6 , _lowercase :Any=1e-2 , _lowercase :int=0.0 , _lowercase :int=0.0 , **_lowercase :Dict , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = image_size lowercase__ = num_channels lowercase__ = patch_sizes lowercase__ = strides lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = mlp_ratios lowercase__ = hidden_act lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = layer_scale_init_value lowercase__ = drop_path_rate lowercase__ = dropout_rate
655
1
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , *_lowercase :Union[str, Any] , **_lowercase :List[Any] ): '''simple docstring''' warnings.warn( "The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use PerceiverImageProcessor instead." , _lowercase , ) super().__init__(*_lowercase , **_lowercase )
655
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 @add_end_docstrings(lowercase_ ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self :Any , *_lowercase :Optional[Any] , **_lowercase :Union[str, Any] ): '''simple docstring''' super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowercase__ = None if self.model.config.prefix is not None: lowercase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowercase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowercase__ , lowercase__ , lowercase__ = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) lowercase__ = {**self._preprocess_params, **preprocess_params} lowercase__ = {**self._forward_params, **forward_params} def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any]=None , _lowercase :List[Any]=None , _lowercase :List[str]=None , _lowercase :Optional[Any]=None , _lowercase :Optional[int]=None , _lowercase :Any=None , _lowercase :Any=None , _lowercase :Dict=None , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ = {} if prefix is not None: lowercase__ = prefix if prefix: lowercase__ = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) lowercase__ = handle_long_generation preprocess_params.update(_lowercase ) lowercase__ = generate_kwargs lowercase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.TENSORS if return_type is not None: lowercase__ = return_type if clean_up_tokenization_spaces is not None: lowercase__ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase ( self :int , *_lowercase :Optional[int] , **_lowercase :List[str] ): '''simple docstring''' if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self :Union[str, Any] , _lowercase :Dict , **_lowercase :Tuple ): '''simple docstring''' return super().__call__(_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple , _lowercase :Optional[int]="" , _lowercase :Tuple=None , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prompt_text if handle_long_generation == "hole": lowercase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: lowercase__ = generate_kwargs["max_new_tokens"] else: lowercase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: lowercase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) lowercase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: lowercase__ = inputs["attention_mask"][:, -keep_length:] return inputs def UpperCAmelCase ( self :str , _lowercase :int , **_lowercase :str ): '''simple docstring''' lowercase__ = model_inputs["input_ids"] lowercase__ = model_inputs.get("attention_mask" , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: lowercase__ = None lowercase__ = None lowercase__ = 1 else: lowercase__ = input_ids.shape[0] lowercase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowercase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: lowercase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: lowercase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowercase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowercase__ = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) lowercase__ = generated_sequence.shape[0] if self.framework == "pt": lowercase__ = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": lowercase__ = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCAmelCase ( self :Any , _lowercase :Tuple , _lowercase :str=ReturnType.FULL_TEXT , _lowercase :Dict=True ): '''simple docstring''' lowercase__ = model_outputs["generated_sequence"][0] lowercase__ = model_outputs["input_ids"] lowercase__ = model_outputs["prompt_text"] lowercase__ = generated_sequence.numpy().tolist() lowercase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowercase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowercase__ = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowercase__ = 0 else: lowercase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: lowercase__ = prompt_text + text[prompt_length:] else: lowercase__ = text[prompt_length:] lowercase__ = {"generated_text": all_text} records.append(_lowercase ) return records
655
1
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def _A ( __magic_name__ ): lowercase__ , lowercase__ = analyze_text(__magic_name__ ) lowercase__ = list(" " + ascii_lowercase ) # what is our total sum of probabilities. lowercase__ = sum(single_char_strings.values() ) # one length string lowercase__ = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowercase__ = single_char_strings[ch] lowercase__ = my_str / all_sum my_fir_sum += prob * math.loga(__magic_name__ ) # entropy formula. # print entropy print(f'''{round(-1 * my_fir_sum ):.1f}''' ) # two len string lowercase__ = sum(two_char_strings.values() ) lowercase__ = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowercase__ = cha + cha if sequence in two_char_strings: lowercase__ = two_char_strings[sequence] lowercase__ = int(__magic_name__ ) / all_sum my_sec_sum += prob * math.loga(__magic_name__ ) # print second entropy print(f'''{round(-1 * my_sec_sum ):.1f}''' ) # print the difference between them print(f'''{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}''' ) def _A ( __magic_name__ ): lowercase__ = Counter() # type: ignore lowercase__ = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(__magic_name__ ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def _A ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
655
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _snake_case = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _snake_case = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def _A ( __magic_name__ ): lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__magic_name__ )[0] @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) lowercase__ = data.reshape(__magic_name__ , __magic_name__ , __magic_name__ , 1 ) return data @deprecated(__magic_name__ , "Please use tf.one_hot on tensors." ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__magic_name__ ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=10 ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(__magic_name__ ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__magic_name__ , __magic_name__ ) return labels class lowerCAmelCase : @deprecated( _lowercase , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Tuple=False , _lowercase :str=False , _lowercase :Dict=dtypes.floataa , _lowercase :Optional[Any]=True , _lowercase :Any=None , ): '''simple docstring''' lowercase__ , lowercase__ = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowercase__ = 1_00_00 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'''images.shape: {images.shape} labels.shape: {labels.shape}''' lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(_lowercase , 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._images @property def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self._labels @property def UpperCAmelCase ( self :Dict ): '''simple docstring''' return self._num_examples @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._epochs_completed def UpperCAmelCase ( self :str , _lowercase :Union[str, Any] , _lowercase :Any=False , _lowercase :Union[str, Any]=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 7_84 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__magic_name__ , "Please write your own downloading logic." ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if not gfile.Exists(__magic_name__ ): gfile.MakeDirs(__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) if not gfile.Exists(__magic_name__ ): urllib.request.urlretrieve(__magic_name__ , __magic_name__ ) # noqa: S310 with gfile.GFile(__magic_name__ ) as f: lowercase__ = f.size() print("Successfully downloaded" , __magic_name__ , __magic_name__ , "bytes." ) return filepath @deprecated( __magic_name__ , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=False , __magic_name__=dtypes.floataa , __magic_name__=True , __magic_name__=5000 , __magic_name__=None , __magic_name__=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__magic_name__ , one_hot=__magic_name__ , dtype=__magic_name__ , seed=__magic_name__ ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = "train-images-idx3-ubyte.gz" lowercase__ = "train-labels-idx1-ubyte.gz" lowercase__ = "t10k-images-idx3-ubyte.gz" lowercase__ = "t10k-labels-idx1-ubyte.gz" lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) if not 0 <= validation_size <= len(__magic_name__ ): lowercase__ = ( "Validation size should be between 0 and " f'''{len(__magic_name__ )}. Received: {validation_size}.''' ) raise ValueError(__magic_name__ ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {"dtype": dtype, "reshape": reshape, "seed": seed} lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ )
655
1
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _snake_case = """.""" # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) _snake_case = [ """Assert""", """AssignVariableOp""", """EmptyTensorList""", """MergeV2Checkpoints""", """ReadVariableOp""", """ResourceGather""", """RestoreV2""", """SaveV2""", """ShardedFilename""", """StatefulPartitionedCall""", """StaticRegexFullMatch""", """VarHandleOp""", ] def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = SavedModel() lowercase__ = [] with open(os.path.join(__magic_name__ , "utils" , "tf_ops" , "onnx.json" ) ) as f: lowercase__ = json.load(__magic_name__ )["opsets"] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ , "rb" ) as f: saved_model.ParseFromString(f.read() ) lowercase__ = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want lowercase__ = sorted(__magic_name__ ) lowercase__ = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(f'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(f'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ , sep="\n" ) else: print(f'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument("""--saved_model_path""", help="""Path of the saved model to check (the .pb file).""") parser.add_argument( """--opset""", default=12, type=int, help="""The ONNX opset against which the model has to be tested.""" ) parser.add_argument( """--framework""", choices=["""onnx"""], default="""onnx""", help="""Frameworks against which to test the saved model.""" ) parser.add_argument( """--strict""", action="""store_true""", help="""Whether make the checking strict (raise errors) or not (raise warnings)""" ) _snake_case = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
655
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
1
import operator as op def _A ( __magic_name__ ): lowercase__ = [] lowercase__ = lambda __magic_name__ , __magic_name__ : int(x / y ) # noqa: E731 integer division operation lowercase__ = { "^": op.pow, "*": op.mul, "/": div, "+": op.add, "-": op.sub, } # operators & their respective operation # print table header print("Symbol".center(8 ) , "Action".center(12 ) , "Stack" , sep=" | " ) print("-" * (30 + len(__magic_name__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(__magic_name__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ("push(" + x + ")").ljust(12 ) , ",".join(__magic_name__ ) , sep=" | " ) else: lowercase__ = stack.pop() # pop stack # output in tabular format print("".rjust(8 ) , ("pop(" + b + ")").ljust(12 ) , ",".join(__magic_name__ ) , sep=" | " ) lowercase__ = stack.pop() # pop stack # output in tabular format print("".rjust(8 ) , ("pop(" + a + ")").ljust(12 ) , ",".join(__magic_name__ ) , sep=" | " ) stack.append( str(opr[x](int(__magic_name__ ) , int(__magic_name__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ("push(" + a + x + b + ")").ljust(12 ) , ",".join(__magic_name__ ) , sep=" | " , ) return int(stack[0] ) if __name__ == "__main__": _snake_case = input("""\n\nEnter a Postfix Equation (space separated) = """).split(""" """) print("""\n\tResult = """, solve(Postfix))
655
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
1
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers _snake_case = [ """python""", """tqdm""", """regex""", """requests""", """packaging""", """filelock""", """numpy""", """tokenizers""", """huggingface-hub""", """safetensors""", """accelerate""", """pyyaml""", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F"""can't find {pkg} in {deps.keys()}, check dependency_versions_table.py""") def _A ( __magic_name__ , __magic_name__=None ): require_version(deps[pkg] , __magic_name__ )
655
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["prompt"] lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] if "image" in inputs: lowercase__ = inputs["image"] else: lowercase__ = None if "mask_image" in inputs: lowercase__ = inputs["mask_image"] else: lowercase__ = None if "original_image" in inputs: lowercase__ = inputs["original_image"] else: lowercase__ = None lowercase__ , lowercase__ = pipe.encode_prompt(_lowercase ) # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_lowercase , _lowercase , _lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowercase , _lowercase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 )
655
1
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) set_seed(770) _snake_case = { """c_attn""": """att_proj""", """c_proj""": """out_proj""", """c_fc""": """in_proj""", """transformer.""": """""", """h.""": """layers.""", """ln_1""": """layernorm_1""", """ln_2""": """layernorm_2""", """ln_f""": """layernorm_final""", """wpe""": """position_embeds_layer""", """wte""": """input_embeds_layer""", } _snake_case = { """text_small""": { """repo_id""": """suno/bark""", """file_name""": """text.pt""", }, """coarse_small""": { """repo_id""": """suno/bark""", """file_name""": """coarse.pt""", }, """fine_small""": { """repo_id""": """suno/bark""", """file_name""": """fine.pt""", }, """text""": { """repo_id""": """suno/bark""", """file_name""": """text_2.pt""", }, """coarse""": { """repo_id""": """suno/bark""", """file_name""": """coarse_2.pt""", }, """fine""": { """repo_id""": """suno/bark""", """file_name""": """fine_2.pt""", }, } _snake_case = os.path.dirname(os.path.abspath(__file__)) _snake_case = os.path.join(os.path.expanduser("""~"""), """.cache""") _snake_case = os.path.join(os.getenv("""XDG_CACHE_HOME""", default_cache_dir), """suno""", """bark_v0""") def _A ( __magic_name__ , __magic_name__=False ): lowercase__ = model_type if use_small: key += "_small" return os.path.join(__magic_name__ , REMOTE_MODEL_PATHS[key]["file_name"] ) def _A ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) hf_hub_download(repo_id=__magic_name__ , filename=__magic_name__ , local_dir=__magic_name__ ) def _A ( __magic_name__ , __magic_name__ , __magic_name__=False , __magic_name__="text" ): if model_type == "text": lowercase__ = BarkSemanticModel lowercase__ = BarkSemanticConfig lowercase__ = BarkSemanticGenerationConfig elif model_type == "coarse": lowercase__ = BarkCoarseModel lowercase__ = BarkCoarseConfig lowercase__ = BarkCoarseGenerationConfig elif model_type == "fine": lowercase__ = BarkFineModel lowercase__ = BarkFineConfig lowercase__ = BarkFineGenerationConfig else: raise NotImplementedError() lowercase__ = f'''{model_type}_small''' if use_small else model_type lowercase__ = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(__magic_name__ ): logger.info(f'''{model_type} model not found, downloading into `{CACHE_DIR}`.''' ) _download(model_info["repo_id"] , model_info["file_name"] ) lowercase__ = torch.load(__magic_name__ , map_location=__magic_name__ ) # this is a hack lowercase__ = checkpoint["model_args"] if "input_vocab_size" not in model_args: lowercase__ = model_args["vocab_size"] lowercase__ = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments lowercase__ = model_args.pop("n_head" ) lowercase__ = model_args.pop("n_embd" ) lowercase__ = model_args.pop("n_layer" ) lowercase__ = ConfigClass(**checkpoint["model_args"] ) lowercase__ = ModelClass(config=__magic_name__ ) lowercase__ = GenerationConfigClass() lowercase__ = model_generation_config lowercase__ = checkpoint["model"] # fixup checkpoint lowercase__ = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(__magic_name__ ): # replace part of the key with corresponding layer name in HF implementation lowercase__ = k[len(__magic_name__ ) :] for old_layer_name in new_layer_name_dict: lowercase__ = new_k.replace(__magic_name__ , new_layer_name_dict[old_layer_name] ) lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = set(state_dict.keys() ) - set(model.state_dict().keys() ) lowercase__ = {k for k in extra_keys if not k.endswith(".attn.bias" )} lowercase__ = set(model.state_dict().keys() ) - set(state_dict.keys() ) lowercase__ = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(__magic_name__ ) != 0: raise ValueError(f'''extra keys found: {extra_keys}''' ) if len(__magic_name__ ) != 0: raise ValueError(f'''missing keys: {missing_keys}''' ) model.load_state_dict(__magic_name__ , strict=__magic_name__ ) lowercase__ = model.num_parameters(exclude_embeddings=__magic_name__ ) lowercase__ = checkpoint["best_val_loss"].item() logger.info(f'''model loaded: {round(n_params/1e6 , 1 )}M params, {round(__magic_name__ , 3 )} loss''' ) model.eval() model.to(__magic_name__ ) del checkpoint, state_dict return model def _A ( __magic_name__ , __magic_name__=False , __magic_name__="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() lowercase__ = "cpu" # do conversion on cpu lowercase__ = _get_ckpt_path(__magic_name__ , use_small=__magic_name__ ) lowercase__ = _load_model(__magic_name__ , __magic_name__ , model_type=__magic_name__ , use_small=__magic_name__ ) # load bark initial model lowercase__ = _bark_load_model(__magic_name__ , "cpu" , model_type=__magic_name__ , use_small=__magic_name__ ) if model_type == "text": lowercase__ = bark_model["model"] if model.num_parameters(exclude_embeddings=__magic_name__ ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model lowercase__ = 5 lowercase__ = 10 if model_type in ["text", "coarse"]: lowercase__ = torch.randint(256 , (batch_size, sequence_length) , dtype=torch.int ) lowercase__ = bark_model(__magic_name__ )[0] lowercase__ = model(__magic_name__ ) # take last logits lowercase__ = output_new_model_total.logits[:, [-1], :] else: lowercase__ = 3 lowercase__ = 8 lowercase__ = torch.randint(256 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) lowercase__ = model(__magic_name__ , __magic_name__ ) lowercase__ = bark_model(__magic_name__ , __magic_name__ ) lowercase__ = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ): lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) lowercase__ = BarkSemanticConfig.from_pretrained(os.path.join(__magic_name__ , "config.json" ) ) lowercase__ = BarkCoarseConfig.from_pretrained(os.path.join(__magic_name__ , "config.json" ) ) lowercase__ = BarkFineConfig.from_pretrained(os.path.join(__magic_name__ , "config.json" ) ) lowercase__ = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) lowercase__ = BarkSemanticModel.from_pretrained(__magic_name__ ) lowercase__ = BarkCoarseModel.from_pretrained(__magic_name__ ) lowercase__ = BarkFineModel.from_pretrained(__magic_name__ ) lowercase__ = EncodecModel.from_pretrained("facebook/encodec_24khz" ) lowercase__ = BarkConfig.from_sub_model_configs( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) lowercase__ = BarkModel(__magic_name__ ) lowercase__ = semantic lowercase__ = coarseAcoustic lowercase__ = fineAcoustic lowercase__ = codec lowercase__ = bark_generation_config Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) bark.save_pretrained(__magic_name__ , repo_id=__magic_name__ , push_to_hub=__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument("""model_type""", type=str, help="""text, coarse or fine.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--is_small""", action="""store_true""", help="""convert the small version instead of the large.""") _snake_case = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
655
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) lowercase__ = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase__ = model(_lowercase )["last_hidden_state"] lowercase__ = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
655
1
from __future__ import annotations from PIL import Image # Define glider example _snake_case = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example _snake_case = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def _A ( __magic_name__ ): lowercase__ = [] for i in range(len(__magic_name__ ) ): lowercase__ = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours lowercase__ = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(__magic_name__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(__magic_name__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(__magic_name__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. lowercase__ = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(__magic_name__ ) return next_generation def _A ( __magic_name__ , __magic_name__ ): lowercase__ = [] for _ in range(__magic_name__ ): # Create output image lowercase__ = Image.new("RGB" , (len(cells[0] ), len(__magic_name__ )) ) lowercase__ = img.load() # Save cells to image for x in range(len(__magic_name__ ) ): for y in range(len(cells[0] ) ): lowercase__ = 255 - cells[y][x] * 255 lowercase__ = (colour, colour, colour) # Save image images.append(__magic_name__ ) lowercase__ = new_generation(__magic_name__ ) return images if __name__ == "__main__": _snake_case = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
655
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
1
import numpy as np def _A ( __magic_name__ , __magic_name__ , __magic_name__ = 1e-12 , __magic_name__ = 100 , ): assert np.shape(__magic_name__ )[0] == np.shape(__magic_name__ )[1] # Ensure proper dimensionality. assert np.shape(__magic_name__ )[0] == np.shape(__magic_name__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(__magic_name__ ) == np.iscomplexobj(__magic_name__ ) lowercase__ = np.iscomplexobj(__magic_name__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(__magic_name__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. lowercase__ = False lowercase__ = 0 lowercase__ = 0 lowercase__ = 1e12 while not convergence: # Multiple matrix by the vector. lowercase__ = np.dot(__magic_name__ , __magic_name__ ) # Normalize the resulting output vector. lowercase__ = w / np.linalg.norm(__magic_name__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) lowercase__ = vector.conj().T if is_complex else vector.T lowercase__ = np.dot(__magic_name__ , np.dot(__magic_name__ , __magic_name__ ) ) # Check convergence. lowercase__ = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: lowercase__ = True lowercase__ = lambda_ if is_complex: lowercase__ = np.real(lambda_ ) return lambda_, vector def _A ( ): lowercase__ = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) lowercase__ = np.array([41, 4, 20] ) lowercase__ = real_input_matrix.astype(np.complexaaa ) lowercase__ = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T lowercase__ = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": lowercase__ = real_input_matrix lowercase__ = real_vector elif problem_type == "complex": lowercase__ = complex_input_matrix lowercase__ = complex_vector # Our implementation. lowercase__ , lowercase__ = power_iteration(__magic_name__ , __magic_name__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). lowercase__ , lowercase__ = np.linalg.eigh(__magic_name__ ) # Last eigenvalue is the maximum one. lowercase__ = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. lowercase__ = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(__magic_name__ ) - np.abs(__magic_name__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
655
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
1
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = SMALL_MODEL_IDENTIFIER lowercase__ = "pt" lowercase__ = "tf" def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :int ): '''simple docstring''' lowercase__ = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = "mock_framework" # Framework provided - return whatever the user provides lowercase__ = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model )
655
1
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": _snake_case = """%20""".join(argv[1:]) if len(argv) > 1 else quote(str(input("""Search: """))) print("""Googling.....""") _snake_case = F"""https://www.google.com/search?q={query}&num=100""" _snake_case = requests.get( url, headers={"""User-Agent""": str(UserAgent().random)}, ) try: _snake_case = ( BeautifulSoup(res.text, """html.parser""") .find("""div""", attrs={"""class""": """yuRUbf"""}) .find("""a""") .get("""href""") ) except AttributeError: _snake_case = parse_qs( BeautifulSoup(res.text, """html.parser""") .find("""div""", attrs={"""class""": """kCrYT"""}) .find("""a""") .get("""href""") )["""url"""][0] webbrowser.open(link)
655
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git_vision_model' def __init__( self :Dict , _lowercase :Dict=7_68 , _lowercase :Dict=30_72 , _lowercase :Tuple=12 , _lowercase :List[str]=12 , _lowercase :Tuple=3 , _lowercase :Dict=2_24 , _lowercase :Tuple=16 , _lowercase :Optional[int]="quick_gelu" , _lowercase :Union[str, Any]=1e-5 , _lowercase :Tuple=0.0 , _lowercase :Tuple=0.02 , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = hidden_size lowercase__ = intermediate_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_channels lowercase__ = patch_size lowercase__ = image_size lowercase__ = initializer_range lowercase__ = attention_dropout lowercase__ = layer_norm_eps lowercase__ = hidden_act @classmethod def UpperCAmelCase ( cls :List[str] , _lowercase :Union[str, os.PathLike] , **_lowercase :Optional[int] ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git' def __init__( self :Union[str, Any] , _lowercase :Dict=None , _lowercase :List[str]=3_05_22 , _lowercase :Tuple=7_68 , _lowercase :Any=6 , _lowercase :Dict=12 , _lowercase :Any=30_72 , _lowercase :List[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :Optional[int]=0.1 , _lowercase :Optional[Any]=10_24 , _lowercase :Any=0.02 , _lowercase :int=1e-12 , _lowercase :List[Any]=0 , _lowercase :int="absolute" , _lowercase :List[str]=True , _lowercase :Any=False , _lowercase :int=1_01 , _lowercase :str=1_02 , _lowercase :Dict=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , pad_token_id=_lowercase , **_lowercase ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowercase__ = GitVisionConfig(**_lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = tie_word_embeddings lowercase__ = num_image_with_embedding lowercase__ = bos_token_id lowercase__ = eos_token_id def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
655
1
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _snake_case = { """facebook/mask2former-swin-small-coco-instance""": ( """https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json""" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'mask2former' __lowerCamelCase = ['swin'] __lowerCamelCase = {'hidden_size': 'hidden_dim'} def __init__( self :str , _lowercase :Optional[Dict] = None , _lowercase :int = 2_56 , _lowercase :int = 2_56 , _lowercase :int = 2_56 , _lowercase :int = 10_24 , _lowercase :str = "relu" , _lowercase :int = 6 , _lowercase :int = 10 , _lowercase :int = 8 , _lowercase :float = 0.0 , _lowercase :int = 20_48 , _lowercase :bool = False , _lowercase :bool = False , _lowercase :int = 4 , _lowercase :int = 2_55 , _lowercase :int = 1_00 , _lowercase :float = 0.1 , _lowercase :float = 2.0 , _lowercase :float = 5.0 , _lowercase :float = 5.0 , _lowercase :int = 1_25_44 , _lowercase :float = 3.0 , _lowercase :float = 0.75 , _lowercase :float = 0.02 , _lowercase :float = 1.0 , _lowercase :bool = True , _lowercase :List[int] = [4, 8, 16, 32] , _lowercase :bool = None , **_lowercase :List[Any] , ): '''simple docstring''' if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `Swin` backbone." ) lowercase__ = CONFIG_MAPPING["swin"]( image_size=2_24 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=_lowercase , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(_lowercase , _lowercase ): lowercase__ = backbone_config.pop("model_type" ) lowercase__ = CONFIG_MAPPING[backbone_model_type] lowercase__ = config_class.from_dict(_lowercase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f'''Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. ''' f'''Supported model types: {','.join(self.backbones_supported )}''' ) lowercase__ = backbone_config lowercase__ = feature_size lowercase__ = mask_feature_size lowercase__ = hidden_dim lowercase__ = encoder_feedforward_dim lowercase__ = activation_function lowercase__ = encoder_layers lowercase__ = decoder_layers lowercase__ = num_attention_heads lowercase__ = dropout lowercase__ = dim_feedforward lowercase__ = pre_norm lowercase__ = enforce_input_projection lowercase__ = common_stride lowercase__ = ignore_value lowercase__ = num_queries lowercase__ = no_object_weight lowercase__ = class_weight lowercase__ = mask_weight lowercase__ = dice_weight lowercase__ = train_num_points lowercase__ = oversample_ratio lowercase__ = importance_sample_ratio lowercase__ = init_std lowercase__ = init_xavier_std lowercase__ = use_auxiliary_loss lowercase__ = feature_strides lowercase__ = output_auxiliary_logits lowercase__ = decoder_layers super().__init__(**_lowercase ) @classmethod def UpperCAmelCase ( cls :int , _lowercase :PretrainedConfig , **_lowercase :Optional[int] ): '''simple docstring''' return cls( backbone_config=_lowercase , **_lowercase , ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.backbone_config.to_dict() lowercase__ = self.__class__.model_type return output
655
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class lowerCAmelCase ( unittest.TestCase ): def __init__( self :int , _lowercase :Dict , _lowercase :Dict=7 , _lowercase :List[str]=3 , _lowercase :str=18 , _lowercase :int=30 , _lowercase :List[str]=4_00 , _lowercase :str=True , _lowercase :Union[str, Any]=None , _lowercase :Dict=True , _lowercase :Union[str, Any]=None , _lowercase :Any=True , _lowercase :Optional[Any]=[0.5, 0.5, 0.5] , _lowercase :List[str]=[0.5, 0.5, 0.5] , ): '''simple docstring''' lowercase__ = size if size is not None else {"shortest_edge": 18} lowercase__ = crop_size if crop_size is not None else {"height": 18, "width": 18} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size lowercase__ = do_center_crop lowercase__ = crop_size lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std def UpperCAmelCase ( self :Any ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = LevitImageProcessor if is_vision_available() else None def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = LevitImageProcessingTester(self ) @property def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , "image_mean" ) ) self.assertTrue(hasattr(_lowercase , "image_std" ) ) self.assertTrue(hasattr(_lowercase , "do_normalize" ) ) self.assertTrue(hasattr(_lowercase , "do_resize" ) ) self.assertTrue(hasattr(_lowercase , "do_center_crop" ) ) self.assertTrue(hasattr(_lowercase , "size" ) ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def UpperCAmelCase ( self :Any ): '''simple docstring''' pass def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowercase__ = image_processing(_lowercase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowercase__ = image_processing(_lowercase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowercase__ = image_processing(_lowercase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
655
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _snake_case = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val def _A ( __magic_name__ ): lowercase__ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowercase__ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) lowercase__ = value else: lowercase__ = value return new_state_dict def _A ( __magic_name__ ): lowercase__ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowercase__ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowercase__ = in_proj_weight_cross_attn[:256, :] lowercase__ = in_proj_bias_cross_attn[:256] lowercase__ = in_proj_weight_cross_attn[256:512, :] lowercase__ = in_proj_bias_cross_attn[256:512] lowercase__ = in_proj_weight_cross_attn[-256:, :] lowercase__ = in_proj_bias_cross_attn[-256:] def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = image.size lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 800 if "detection" in checkpoint_url else 1000 lowercase__ = target_max_size / current_max_size lowercase__ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _A ( __magic_name__ ): lowercase__ = F.to_tensor(__magic_name__ ) lowercase__ = F.normalize(__magic_name__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): logger.info("Converting model..." ) # load original state dict lowercase__ = torch.hub.load_state_dict_from_url(__magic_name__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowercase__ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val # create HuggingFace model and load state dict lowercase__ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: lowercase__ = 15 lowercase__ = 2 lowercase__ = {0: "table", 1: "table rotated"} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 125 lowercase__ = 6 lowercase__ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) lowercase__ = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion lowercase__ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" lowercase__ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=__magic_name__ ) lowercase__ = Image.open(__magic_name__ ).convert("RGB" ) lowercase__ = normalize(resize(__magic_name__ , __magic_name__ ) ).unsqueeze(0 ) lowercase__ = model(__magic_name__ ) if "detection" in checkpoint_url: lowercase__ = (1, 15, 3) lowercase__ = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) lowercase__ = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: lowercase__ = (1, 125, 7) lowercase__ = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) lowercase__ = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __magic_name__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) lowercase__ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _snake_case = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
655
1
from collections import defaultdict from math import ceil, sqrt def _A ( __magic_name__ = 100_0000 , __magic_name__ = 10 ): lowercase__ = defaultdict(__magic_name__ ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: lowercase__ = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: lowercase__ = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(__magic_name__ , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(F"""{solution() = }""")
655
from typing import TYPE_CHECKING from ...utils import _LazyModule _snake_case = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
655
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = """ Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _A ( __magic_name__ , __magic_name__ , __magic_name__=8 ): lowercase__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :UNetaDConditionModel , _lowercase :DDPMScheduler , _lowercase :VQModel , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) lowercase__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :int , _lowercase :str ): '''simple docstring''' if latents is None: lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase__ = latents.to(_lowercase ) lowercase__ = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self :int , _lowercase :int=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) lowercase__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase__ = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase__ , lowercase__ = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. lowercase__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__( self :int , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 1_00 , _lowercase :float = 4.0 , _lowercase :int = 1 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = self._execution_device lowercase__ = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) lowercase__ = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: lowercase__ = image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) lowercase__ = self.scheduler.timesteps lowercase__ = self.unet.config.in_channels lowercase__ , lowercase__ = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) # create initial latent lowercase__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowercase , _lowercase , _lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance lowercase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase__ = {"image_embeds": image_embeds} lowercase__ = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) lowercase__ , lowercase__ = noise_pred.chunk(2 ) lowercase__ , lowercase__ = variance_pred.chunk(2 ) lowercase__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase__ = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing lowercase__ = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase__ = image * 0.5 + 0.5 lowercase__ = image.clamp(0 , 1 ) lowercase__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
655
1
def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): return round(float(moles / volume ) * nfactor ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): return round(float((moles * 0.0_821 * temperature) / (volume) ) ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): return round(float((moles * 0.0_821 * temperature) / (pressure) ) ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): return round(float((pressure * volume) / (0.0_821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
655
import inspect import unittest class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :int ): '''simple docstring''' try: import diffusers # noqa: F401 except ImportError: assert False def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' import diffusers from diffusers.dependency_versions_table import deps lowercase__ = inspect.getmembers(_lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowercase__ = "k-diffusion" elif backend == "invisible_watermark": lowercase__ = "invisible-watermark" assert backend in deps, f'''{backend} is not in the deps table!'''
655
1
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _snake_case = { """E""": 12.70, """T""": 9.06, """A""": 8.17, """O""": 7.51, """I""": 6.97, """N""": 6.75, """S""": 6.33, """H""": 6.09, """R""": 5.99, """D""": 4.25, """L""": 4.03, """C""": 2.78, """U""": 2.76, """M""": 2.41, """W""": 2.36, """F""": 2.23, """G""": 2.02, """Y""": 1.97, """P""": 1.93, """B""": 1.29, """V""": 0.98, """K""": 0.77, """J""": 0.15, """X""": 0.15, """Q""": 0.10, """Z""": 0.07, } _snake_case = """ETAOINSHRDLCUMWFGYPBVKJXQZ""" _snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" def _A ( __magic_name__ ): lowercase__ = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def _A ( __magic_name__ ): return x[0] def _A ( __magic_name__ ): lowercase__ = get_letter_count(__magic_name__ ) lowercase__ = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__magic_name__ ) lowercase__ = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=__magic_name__ ) lowercase__ = "".join(freq_to_letter[freq] ) lowercase__ = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__magic_name__ , reverse=__magic_name__ ) lowercase__ = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__magic_name__ ) def _A ( __magic_name__ ): lowercase__ = get_frequency_order(__magic_name__ ) lowercase__ = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
655
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 # setable values __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None @classmethod def UpperCAmelCase ( cls :Union[str, Any] , _lowercase :CommonSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray ): '''simple docstring''' return cls(common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase ) @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCamelCase = 42 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return True @register_to_config def __init__( self :str , _lowercase :int = 10_00 , _lowercase :float = 0.0001 , _lowercase :float = 0.02 , _lowercase :str = "linear" , _lowercase :Optional[jnp.ndarray] = None , _lowercase :str = "fixed_small" , _lowercase :bool = True , _lowercase :str = "epsilon" , _lowercase :jnp.dtype = jnp.floataa , ): '''simple docstring''' lowercase__ = dtype def UpperCAmelCase ( self :str , _lowercase :Optional[CommonSchedulerState] = None ): '''simple docstring''' if common is None: lowercase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowercase__ = jnp.array(1.0 , dtype=self.dtype ) lowercase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase ( self :List[str] , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :Tuple = () ): '''simple docstring''' lowercase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowercase__ = (jnp.arange(0 , _lowercase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Tuple , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :List[str]=None , _lowercase :Tuple=None ): '''simple docstring''' lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowercase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowercase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowercase__ = jnp.clip(_lowercase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowercase__ = jnp.log(jnp.clip(_lowercase , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowercase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowercase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowercase__ = variance lowercase__ = state.common.betas[t] lowercase__ = (predicted_variance + 1) / 2 lowercase__ = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase ( self :Optional[int] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :int , _lowercase :jnp.ndarray , _lowercase :Optional[jax.random.KeyArray] = None , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = timestep if key is None: lowercase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ = jnp.split(_lowercase , sample.shape[1] , axis=1 ) else: lowercase__ = None # 1. compute alphas, betas lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowercase__ = 1 - alpha_prod_t lowercase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowercase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowercase__ = model_output elif self.config.prediction_type == "v_prediction": lowercase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowercase__ = jnp.clip(_lowercase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowercase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowercase__ = jax.random.split(_lowercase , num=1 ) lowercase__ = jax.random.normal(_lowercase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowercase , _lowercase , predicted_variance=_lowercase ) ** 0.5) * noise lowercase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowercase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowercase , state=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return add_noise_common(state.common , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return get_velocity_common(state.common , _lowercase , _lowercase , _lowercase ) def __len__( self :List[str] ): '''simple docstring''' return self.config.num_train_timesteps
655
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { """configuration_mgp_str""": ["""MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MgpstrConfig"""], """processing_mgp_str""": ["""MgpstrProcessor"""], """tokenization_mgp_str""": ["""MgpstrTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST""", """MgpstrModel""", """MgpstrPreTrainedModel""", """MgpstrForSceneTextRecognition""", ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _snake_case = logging.get_logger(__name__) _snake_case = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _snake_case = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCAmelCase : __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Model type selected in the list: ' + ', '.join(lowercase_ )} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) __lowerCamelCase = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) __lowerCamelCase = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) __lowerCamelCase = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) __lowerCamelCase = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'train' __lowerCamelCase = 'dev' class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 def __init__( self :Optional[Any] , _lowercase :SquadDataTrainingArguments , _lowercase :PreTrainedTokenizer , _lowercase :Optional[int] = None , _lowercase :Union[str, Split] = Split.train , _lowercase :Optional[bool] = False , _lowercase :Optional[str] = None , _lowercase :Optional[str] = "pt" , ): '''simple docstring''' lowercase__ = args lowercase__ = is_language_sensitive lowercase__ = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_lowercase , _lowercase ): try: lowercase__ = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) lowercase__ = mode # Load data features from cache or dataset file lowercase__ = "v2" if args.version_2_with_negative else "v1" lowercase__ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase__ = cached_features_file + ".lock" with FileLock(_lowercase ): if os.path.exists(_lowercase ) and not args.overwrite_cache: lowercase__ = time.time() lowercase__ = torch.load(_lowercase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowercase__ = self.old_features["features"] lowercase__ = self.old_features.get("dataset" , _lowercase ) lowercase__ = self.old_features.get("examples" , _lowercase ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: lowercase__ = self.processor.get_dev_examples(args.data_dir ) else: lowercase__ = self.processor.get_train_examples(args.data_dir ) lowercase__ , lowercase__ = squad_convert_examples_to_features( examples=self.examples , tokenizer=_lowercase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_lowercase , ) lowercase__ = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _lowercase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self :Dict ): '''simple docstring''' return len(self.features ) def __getitem__( self :Any , _lowercase :Any ): '''simple docstring''' lowercase__ = self.features[i] lowercase__ = torch.tensor(feature.input_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.attention_mask , dtype=torch.long ) lowercase__ = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.cls_index , dtype=torch.long ) lowercase__ = torch.tensor(feature.p_mask , dtype=torch.float ) lowercase__ = torch.tensor(feature.is_impossible , dtype=torch.float ) lowercase__ = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowercase__ = torch.tensor(feature.start_position , dtype=torch.long ) lowercase__ = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
655
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = """▁""" _snake_case = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} _snake_case = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } _snake_case = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } _snake_case = { """ernie-m-base""": 514, """ernie-m-large""": 514, } _snake_case = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ["input_ids"] __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = RESOURCE_FILES_NAMES def __init__( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=False , _lowercase :Dict="utf8" , _lowercase :Optional[Any]="[UNK]" , _lowercase :Optional[int]="[SEP]" , _lowercase :List[str]="[PAD]" , _lowercase :Dict="[CLS]" , _lowercase :Optional[Any]="[MASK]" , _lowercase :Optional[Dict[str, Any]] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , vocab_file=_lowercase , encoding=_lowercase , sp_model_kwargs=self.sp_model_kwargs , **_lowercase , ) lowercase__ = do_lower_case lowercase__ = sentencepiece_model_ckpt lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase__ = self.load_vocab(filepath=_lowercase ) else: lowercase__ = {self.sp_model.id_to_piece(_lowercase ): id for id in range(self.sp_model.get_piece_size() )} lowercase__ = {v: k for k, v in self.vocab.items()} def UpperCAmelCase ( self :Any , _lowercase :Dict ): '''simple docstring''' if text is None: return None lowercase__ = self.tokenize(_lowercase ) lowercase__ , lowercase__ = "", [] for i, ch in enumerate(_lowercase ): if ch in self.SP_CHAR_MAPPING: lowercase__ = self.SP_CHAR_MAPPING.get(_lowercase ) else: lowercase__ = unicodedata.normalize("NFKC" , _lowercase ) if self.is_whitespace(_lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(_lowercase ) ) lowercase__ , lowercase__ , lowercase__ = normalized_text, [], 0 if self.do_lower_case: lowercase__ = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase__ = token[1:] lowercase__ = text[offset:].index(_lowercase ) + offset lowercase__ = start + len(_lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase__ = end return token_mapping @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return len(self.vocab ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self :Any ): '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self :Optional[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCAmelCase ( self :List[str] , _lowercase :Optional[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(_lowercase , _lowercase ) for c in text) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any]=False , _lowercase :Optional[int]=64 , _lowercase :Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get("enable_sampling" ) is True: lowercase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: lowercase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: lowercase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: lowercase__ = self.sp_model.EncodeAsPieces(_lowercase ) else: lowercase__ = self.sp_model.SampleEncodeAsPieces(_lowercase , _lowercase , _lowercase ) lowercase__ = [] for pi, piece in enumerate(_lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(_lowercase ) and pi != 0: new_pieces.append(_lowercase ) continue else: continue lowercase__ = 0 for i, chunk in enumerate(_lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(_lowercase ) or self.is_punct(_lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(_lowercase ) lowercase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i if len(_lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCAmelCase ( self :Tuple , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = self.convert_ids_to_tokens(_lowercase ) lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' return self.vocab.get(_lowercase , self.vocab.get(self.unk_token ) ) def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' return self.reverse_vocab.get(_lowercase , self.unk_token ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :Tuple=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCAmelCase ( self :Dict , _lowercase :int , _lowercase :Union[str, Any]=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Dict=None , _lowercase :Optional[Any]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(_lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(_lowercase ) + 1) + [1] * (len(_lowercase ) + 3) def UpperCAmelCase ( self :str , _lowercase :Optional[int] ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCAmelCase ( self :int , _lowercase :Dict ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCAmelCase ( self :List[str] , _lowercase :List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(_lowercase ) == 1: lowercase__ = unicodedata.category(_lowercase ) if cat == "Zs": return True return False def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = {} with io.open(_lowercase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(_lowercase ): lowercase__ = line.rstrip("\n" ) lowercase__ = int(_lowercase ) return token_to_idx def UpperCAmelCase ( self :List[str] , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' lowercase__ = 0 if os.path.isdir(_lowercase ): lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowercase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(_lowercase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowercase__ = token_index writer.write(token + "\n" ) index += 1 lowercase__ = os.path.join(_lowercase , "sentencepiece.bpe.model" ) with open(_lowercase , "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(_lowercase ) return (vocab_file,)
655
1
from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer import PriorTransformer from .ta_film_transformer import TaFilmDecoder from .transformer_ad import TransformeraDModel from .unet_ad import UNetaDModel from .unet_ad import UNetaDModel from .unet_ad_condition import UNetaDConditionModel from .unet_ad_condition import UNetaDConditionModel from .vq_model import VQModel if is_flax_available(): from .controlnet_flax import FlaxControlNetModel from .unet_ad_condition_flax import FlaxUNetaDConditionModel from .vae_flax import FlaxAutoencoderKL
655
def _A ( __magic_name__ ): lowercase__ = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def _A ( __magic_name__ = 100 ): lowercase__ = 1 lowercase__ = 2 for i in range(2 , max_n + 1 ): lowercase__ = pre_numerator lowercase__ = 2 * i // 3 if i % 3 == 0 else 1 lowercase__ = cur_numerator lowercase__ = e_cont * pre_numerator + temp return sum_digits(__magic_name__ ) if __name__ == "__main__": print(F"""{solution() = }""")
655
1
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class lowerCAmelCase : __lowerCamelCase = 42 # [batch_size x 3] __lowerCamelCase = 42 # [batch_size x 3] __lowerCamelCase = 42 # [batch_size x 3] __lowerCamelCase = 42 # [batch_size x 3] __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = torch.arange(self.height * self.width ) lowercase__ = torch.stack( [ pixel_indices % self.width, torch.div(_lowercase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , *lowercase__ = self.shape lowercase__ = int(np.prod(_lowercase ) ) lowercase__ = self.get_image_coords() lowercase__ = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) lowercase__ = self.get_camera_rays(_lowercase ) lowercase__ = rays.view(_lowercase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCAmelCase ( self :Union[str, Any] , _lowercase :torch.Tensor ): '''simple docstring''' lowercase__ , *lowercase__ , lowercase__ = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] lowercase__ = coords.view(_lowercase , -1 , 2 ) lowercase__ = self.resolution() lowercase__ = self.fov() lowercase__ = (flat.float() / (res - 1)) * 2 - 1 lowercase__ = fracs * torch.tan(fov / 2 ) lowercase__ = fracs.view(_lowercase , -1 , 2 ) lowercase__ = ( self.z.view(_lowercase , 1 , 3 ) + self.x.view(_lowercase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(_lowercase , 1 , 3 ) * fracs[:, :, 1:] ) lowercase__ = directions / directions.norm(dim=-1 , keepdim=_lowercase ) lowercase__ = torch.stack( [ torch.broadcast_to(self.origin.view(_lowercase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(_lowercase , *_lowercase , 2 , 3 ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :int , _lowercase :int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=_lowercase , height=_lowercase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _A ( __magic_name__ ): lowercase__ = [] lowercase__ = [] lowercase__ = [] lowercase__ = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): lowercase__ = np.array([np.sin(__magic_name__ ), np.cos(__magic_name__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) lowercase__ = -z * 4 lowercase__ = np.array([np.cos(__magic_name__ ), -np.sin(__magic_name__ ), 0.0] ) lowercase__ = np.cross(__magic_name__ , __magic_name__ ) origins.append(__magic_name__ ) xs.append(__magic_name__ ) ys.append(__magic_name__ ) zs.append(__magic_name__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(__magic_name__ , axis=0 ) ).float() , x=torch.from_numpy(np.stack(__magic_name__ , axis=0 ) ).float() , y=torch.from_numpy(np.stack(__magic_name__ , axis=0 ) ).float() , z=torch.from_numpy(np.stack(__magic_name__ , axis=0 ) ).float() , width=__magic_name__ , height=__magic_name__ , x_fov=0.7 , y_fov=0.7 , shape=(1, len(__magic_name__ )) , )
655
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'AutoTokenizer' __lowerCamelCase = ['tokenizer'] __lowerCamelCase = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self :Dict , _lowercase :List[str] , _lowercase :List[Any]=None ): '''simple docstring''' super().__init__(_lowercase ) lowercase__ = speaker_embeddings @classmethod def UpperCAmelCase ( cls :Any , _lowercase :int , _lowercase :str="speaker_embeddings_path.json" , **_lowercase :List[str] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase__ = get_file_from_repo( _lowercase , _lowercase , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowercase , _lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase__ = None else: with open(_lowercase ) as speaker_embeddings_json: lowercase__ = json.load(_lowercase ) else: lowercase__ = None lowercase__ = AutoTokenizer.from_pretrained(_lowercase , **_lowercase ) return cls(tokenizer=_lowercase , speaker_embeddings=_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :List[str]="speaker_embeddings_path.json" , _lowercase :Any="speaker_embeddings" , _lowercase :bool = False , **_lowercase :Any , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowercase , _lowercase , "v2" ) , exist_ok=_lowercase ) lowercase__ = {} lowercase__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase__ = self._load_voice_preset(_lowercase ) lowercase__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowercase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowercase , ) lowercase__ = os.path.join(_lowercase , f'''{prompt_key}_{key}.npy''' ) lowercase__ = tmp_dict with open(os.path.join(_lowercase , _lowercase ) , "w" ) as fp: json.dump(_lowercase , _lowercase ) super().save_pretrained(_lowercase , _lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :str = None , **_lowercase :List[Any] ): '''simple docstring''' lowercase__ = self.speaker_embeddings[voice_preset] lowercase__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase__ = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) lowercase__ = np.load(_lowercase ) return voice_preset_dict def UpperCAmelCase ( self :Optional[int] , _lowercase :Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Optional[Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=None , _lowercase :List[str]="pt" , _lowercase :List[Any]=2_56 , _lowercase :List[str]=False , _lowercase :Union[str, Any]=True , _lowercase :Dict=False , **_lowercase :Tuple , ): '''simple docstring''' if voice_preset is not None and not isinstance(_lowercase , _lowercase ): if ( isinstance(_lowercase , _lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase__ = self._load_voice_preset(_lowercase ) else: if isinstance(_lowercase , _lowercase ) and not voice_preset.endswith(".npz" ): lowercase__ = voice_preset + ".npz" lowercase__ = np.load(_lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowercase , **_lowercase ) lowercase__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) lowercase__ = self.tokenizer( _lowercase , return_tensors=_lowercase , padding="max_length" , max_length=_lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , add_special_tokens=_lowercase , **_lowercase , ) if voice_preset is not None: lowercase__ = voice_preset return encoded_text
655
1
from __future__ import annotations def _A ( __magic_name__ ): lowercase__ = len(__magic_name__ ) # We need to create solution object to save path. lowercase__ = [[0 for _ in range(__magic_name__ )] for _ in range(__magic_name__ )] lowercase__ = run_maze(__magic_name__ , 0 , 0 , __magic_name__ ) if solved: print("\n".join(str(__magic_name__ ) for row in solutions ) ) else: print("No solution exists!" ) return solved def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = len(__magic_name__ ) # Final check point. if i == j == (size - 1): lowercase__ = 1 return True lowercase__ = (not i < 0) and (not j < 0) # Check lower bounds lowercase__ = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. lowercase__ = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited lowercase__ = 1 # check for directions if ( run_maze(__magic_name__ , i + 1 , __magic_name__ , __magic_name__ ) or run_maze(__magic_name__ , __magic_name__ , j + 1 , __magic_name__ ) or run_maze(__magic_name__ , i - 1 , __magic_name__ , __magic_name__ ) or run_maze(__magic_name__ , __magic_name__ , j - 1 , __magic_name__ ) ): return True lowercase__ = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
655
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__magic_name__ ): # Forward propagation lowercase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? lowercase__ = (expected / 100) - layer_a # Error delta lowercase__ = layer_1_error * sigmoid_function(__magic_name__ , __magic_name__ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input("""Expected value: """)) _snake_case = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
655
1
from sklearn.metrics import recall_score import datasets _snake_case = """ Recall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation: Recall = TP / (TP + FN) Where TP is the true positives and FN is the false negatives. """ _snake_case = """ Args: - **predictions** (`list` of `int`): The predicted labels. - **references** (`list` of `int`): The ground truth labels. - **labels** (`list` of `int`): The set of labels to include when `average` is not set to `binary`, and their order when average is `None`. Labels present in the data can be excluded in this input, for example to calculate a multiclass average ignoring a majority negative class, while labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in y_true and y_pred are used in sorted order. Defaults to None. - **pos_label** (`int`): The class label to use as the 'positive class' when calculating the recall. Defaults to `1`. - **average** (`string`): This parameter is required for multiclass/multilabel targets. If None, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`. - `'binary'`: Only report results for the class specified by `pos_label`. This is applicable only if the target labels and predictions are binary. - `'micro'`: Calculate metrics globally by counting the total true positives, false negatives, and false positives. - `'macro'`: Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - `'weighted'`: Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. Note that it can result in an F-score that is not between precision and recall. - `'samples'`: Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). - **sample_weight** (`list` of `float`): Sample weights Defaults to `None`. - **zero_division** (): Sets the value to return when there is a zero division. Defaults to . - `'warn'`: If there is a zero division, the return value is `0`, but warnings are also raised. - `0`: If there is a zero division, the return value is `0`. - `1`: If there is a zero division, the return value is `1`. Returns: - **recall** (`float`, or `array` of `float`): Either the general recall score, or the recall scores for individual classes, depending on the values input to `labels` and `average`. Minimum possible value is 0. Maximum possible value is 1. A higher recall means that more of the positive examples have been labeled correctly. Therefore, a higher recall is generally considered better. Examples: Example 1-A simple example with some errors >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1]) >>> print(results) {'recall': 0.6666666666666666} Example 2-The same example as Example 1, but with `pos_label=0` instead of the default `pos_label=1`. >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], pos_label=0) >>> print(results) {'recall': 0.5} Example 3-The same example as Example 1, but with `sample_weight` included. >>> recall_metric = datasets.load_metric('recall') >>> sample_weight = [0.9, 0.2, 0.9, 0.3, 0.8] >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], sample_weight=sample_weight) >>> print(results) {'recall': 0.55} Example 4-A multiclass example, using different averages. >>> recall_metric = datasets.load_metric('recall') >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = recall_metric.compute(predictions=predictions, references=references, average='macro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='micro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='weighted') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {'recall': array([1., 0., 0.])} """ _snake_case = """ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("int32" ) ), "references": datasets.Sequence(datasets.Value("int32" ) ), } if self.config_name == "multilabel" else { "predictions": datasets.Value("int32" ), "references": datasets.Value("int32" ), } ) , reference_urls=["https://scikit-learn.org/stable/modules/generated/sklearn.metrics.recall_score.html"] , ) def UpperCAmelCase ( self :str , _lowercase :List[Any] , _lowercase :Dict , _lowercase :Optional[int]=None , _lowercase :Tuple=1 , _lowercase :str="binary" , _lowercase :Any=None , _lowercase :Tuple="warn" , ): '''simple docstring''' lowercase__ = recall_score( _lowercase , _lowercase , labels=_lowercase , pos_label=_lowercase , average=_lowercase , sample_weight=_lowercase , zero_division=_lowercase , ) return {"recall": float(_lowercase ) if score.size == 1 else score}
655
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'van' def __init__( self :Optional[Any] , _lowercase :Dict=2_24 , _lowercase :Union[str, Any]=3 , _lowercase :List[Any]=[7, 3, 3, 3] , _lowercase :Any=[4, 2, 2, 2] , _lowercase :Union[str, Any]=[64, 1_28, 3_20, 5_12] , _lowercase :List[Any]=[3, 3, 12, 3] , _lowercase :Dict=[8, 8, 4, 4] , _lowercase :int="gelu" , _lowercase :List[Any]=0.02 , _lowercase :List[Any]=1e-6 , _lowercase :Any=1e-2 , _lowercase :int=0.0 , _lowercase :int=0.0 , **_lowercase :Dict , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = image_size lowercase__ = num_channels lowercase__ = patch_sizes lowercase__ = strides lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = mlp_ratios lowercase__ = hidden_act lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = layer_scale_init_value lowercase__ = drop_path_rate lowercase__ = dropout_rate
655
1
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = RoCBertTokenizer __lowerCamelCase = None __lowerCamelCase = False __lowerCamelCase = True __lowerCamelCase = filter_non_english def UpperCAmelCase ( self :str ): '''simple docstring''' super().setUp() lowercase__ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "你", "好", "是", "谁", "a", "b", "c", "d"] lowercase__ = {} lowercase__ = {} for i, value in enumerate(_lowercase ): lowercase__ = i lowercase__ = i lowercase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_shape_file"] ) lowercase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_pronunciation_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) with open(self.word_shape_file , "w" , encoding="utf-8" ) as word_shape_writer: json.dump(_lowercase , _lowercase , ensure_ascii=_lowercase ) with open(self.word_pronunciation_file , "w" , encoding="utf-8" ) as word_pronunciation_writer: json.dump(_lowercase , _lowercase , ensure_ascii=_lowercase ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) lowercase__ = tokenizer.tokenize("你好[SEP]你是谁" ) self.assertListEqual(_lowercase , ["你", "好", "[SEP]", "你", "是", "谁"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = RoCBertBasicTokenizer(do_lower_case=_lowercase , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] lowercase__ = {} for i, token in enumerate(_lowercase ): lowercase__ = i lowercase__ = RoCBertWordpieceTokenizer(vocab=_lowercase , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def UpperCAmelCase ( self :int ): '''simple docstring''' self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowercase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) if self.test_rust_tokenizer: lowercase__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(_lowercase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) lowercase__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' lowercase__ = tokenizer_r.encode_plus( _lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , return_offsets_mapping=_lowercase , add_special_tokens=_lowercase , ) lowercase__ = tokenizer_r.do_lower_case if hasattr(_lowercase , "do_lower_case" ) else False lowercase__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = ["的", "人", "有"] lowercase__ = "".join(_lowercase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase__ = True lowercase__ = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) lowercase__ = tokenizer_p.encode(_lowercase , add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.encode(_lowercase , add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.convert_ids_to_tokens(_lowercase ) lowercase__ = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowercase , _lowercase ) self.assertListEqual(_lowercase , _lowercase ) lowercase__ = False lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) lowercase__ = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) lowercase__ = tokenizer_r.encode(_lowercase , add_special_tokens=_lowercase ) lowercase__ = tokenizer_p.encode(_lowercase , add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.convert_ids_to_tokens(_lowercase ) lowercase__ = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that only the first Chinese character is not preceded by "##". lowercase__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(_lowercase ) ] self.assertListEqual(_lowercase , _lowercase ) self.assertListEqual(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) lowercase__ = tokenizer.encode("你好" , add_special_tokens=_lowercase ) lowercase__ = tokenizer.encode("你是谁" , add_special_tokens=_lowercase ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowercase ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowercase , _lowercase ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.get_tokenizers(do_lower_case=_lowercase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): lowercase__ = "你好,你是谁" lowercase__ = tokenizer.tokenize(_lowercase ) lowercase__ = tokenizer.convert_tokens_to_ids(_lowercase ) lowercase__ = tokenizer.convert_tokens_to_shape_ids(_lowercase ) lowercase__ = tokenizer.convert_tokens_to_pronunciation_ids(_lowercase ) lowercase__ = tokenizer.prepare_for_model( _lowercase , _lowercase , _lowercase , add_special_tokens=_lowercase ) lowercase__ = tokenizer.encode_plus(_lowercase , add_special_tokens=_lowercase ) self.assertEqual(_lowercase , _lowercase )
655
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 @add_end_docstrings(lowercase_ ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self :Any , *_lowercase :Optional[Any] , **_lowercase :Union[str, Any] ): '''simple docstring''' super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowercase__ = None if self.model.config.prefix is not None: lowercase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowercase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowercase__ , lowercase__ , lowercase__ = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) lowercase__ = {**self._preprocess_params, **preprocess_params} lowercase__ = {**self._forward_params, **forward_params} def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any]=None , _lowercase :List[Any]=None , _lowercase :List[str]=None , _lowercase :Optional[Any]=None , _lowercase :Optional[int]=None , _lowercase :Any=None , _lowercase :Any=None , _lowercase :Dict=None , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ = {} if prefix is not None: lowercase__ = prefix if prefix: lowercase__ = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) lowercase__ = handle_long_generation preprocess_params.update(_lowercase ) lowercase__ = generate_kwargs lowercase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.TENSORS if return_type is not None: lowercase__ = return_type if clean_up_tokenization_spaces is not None: lowercase__ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase ( self :int , *_lowercase :Optional[int] , **_lowercase :List[str] ): '''simple docstring''' if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self :Union[str, Any] , _lowercase :Dict , **_lowercase :Tuple ): '''simple docstring''' return super().__call__(_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple , _lowercase :Optional[int]="" , _lowercase :Tuple=None , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prompt_text if handle_long_generation == "hole": lowercase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: lowercase__ = generate_kwargs["max_new_tokens"] else: lowercase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: lowercase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) lowercase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: lowercase__ = inputs["attention_mask"][:, -keep_length:] return inputs def UpperCAmelCase ( self :str , _lowercase :int , **_lowercase :str ): '''simple docstring''' lowercase__ = model_inputs["input_ids"] lowercase__ = model_inputs.get("attention_mask" , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: lowercase__ = None lowercase__ = None lowercase__ = 1 else: lowercase__ = input_ids.shape[0] lowercase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowercase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: lowercase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: lowercase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowercase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowercase__ = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) lowercase__ = generated_sequence.shape[0] if self.framework == "pt": lowercase__ = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": lowercase__ = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCAmelCase ( self :Any , _lowercase :Tuple , _lowercase :str=ReturnType.FULL_TEXT , _lowercase :Dict=True ): '''simple docstring''' lowercase__ = model_outputs["generated_sequence"][0] lowercase__ = model_outputs["input_ids"] lowercase__ = model_outputs["prompt_text"] lowercase__ = generated_sequence.numpy().tolist() lowercase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowercase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowercase__ = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowercase__ = 0 else: lowercase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: lowercase__ = prompt_text + text[prompt_length:] else: lowercase__ = text[prompt_length:] lowercase__ = {"generated_text": all_text} records.append(_lowercase ) return records
655
1
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask _snake_case = logging.getLogger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'token-classification' def __init__( self :int , _lowercase :Union[str, Any] ): '''simple docstring''' if type(_lowercase ) == dict: lowercase__ = Namespace(**_lowercase ) lowercase__ = import_module("tasks" ) try: lowercase__ = getattr(_lowercase , hparams.task_type ) lowercase__ = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) lowercase__ = self.token_classification_task.get_labels(hparams.labels ) lowercase__ = CrossEntropyLoss().ignore_index super().__init__(_lowercase , len(self.labels ) , self.mode ) def UpperCAmelCase ( self :str , **_lowercase :Any ): '''simple docstring''' return self.model(**_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :List[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": lowercase__ = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids lowercase__ = self(**_lowercase ) lowercase__ = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = self.hparams for mode in ["train", "dev", "test"]: lowercase__ = self._feature_file(_lowercase ) if os.path.exists(_lowercase ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , _lowercase ) lowercase__ = torch.load(_lowercase ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) lowercase__ = self.token_classification_task.read_examples_from_file(args.data_dir , _lowercase ) lowercase__ = self.token_classification_task.convert_examples_to_features( _lowercase , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=_lowercase , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , _lowercase ) torch.save(_lowercase , _lowercase ) def UpperCAmelCase ( self :int , _lowercase :int , _lowercase :int , _lowercase :bool = False ): '''simple docstring''' lowercase__ = self._feature_file(_lowercase ) logger.info("Loading features from cached file %s" , _lowercase ) lowercase__ = torch.load(_lowercase ) lowercase__ = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) lowercase__ = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: lowercase__ = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: lowercase__ = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) lowercase__ = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(_lowercase , _lowercase , _lowercase , _lowercase ) , batch_size=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :Any , _lowercase :Optional[Any] ): '''simple docstring''' """Compute validation""" "" lowercase__ = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": lowercase__ = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids lowercase__ = self(**_lowercase ) lowercase__ , lowercase__ = outputs[:2] lowercase__ = logits.detach().cpu().numpy() lowercase__ = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCAmelCase ( self :List[str] , _lowercase :Dict ): '''simple docstring''' lowercase__ = torch.stack([x["val_loss"] for x in outputs] ).mean() lowercase__ = np.concatenate([x["pred"] for x in outputs] , axis=0 ) lowercase__ = np.argmax(_lowercase , axis=2 ) lowercase__ = np.concatenate([x["target"] for x in outputs] , axis=0 ) lowercase__ = dict(enumerate(self.labels ) ) lowercase__ = [[] for _ in range(out_label_ids.shape[0] )] lowercase__ = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) lowercase__ = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(_lowercase , _lowercase ), "precision": precision_score(_lowercase , _lowercase ), "recall": recall_score(_lowercase , _lowercase ), "f1": fa_score(_lowercase , _lowercase ), } lowercase__ = dict(results.items() ) lowercase__ = results return ret, preds_list, out_label_list def UpperCAmelCase ( self :List[Any] , _lowercase :List[str] ): '''simple docstring''' lowercase__ , lowercase__ , lowercase__ = self._eval_end(_lowercase ) lowercase__ = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCAmelCase ( self :Any , _lowercase :Any ): '''simple docstring''' lowercase__ , lowercase__ , lowercase__ = self._eval_end(_lowercase ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 lowercase__ = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCAmelCase ( _lowercase :Any , _lowercase :Any ): '''simple docstring''' BaseTransformer.add_model_specific_args(_lowercase , _lowercase ) parser.add_argument( "--task_type" , default="NER" , type=_lowercase , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=1_28 , type=_lowercase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=_lowercase , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=_lowercase , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": _snake_case = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) _snake_case = NERTransformer.add_model_specific_args(parser, os.getcwd()) _snake_case = parser.parse_args() _snake_case = NERTransformer(args) _snake_case = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 _snake_case = sorted(glob.glob(os.path.join(args.output_dir, """checkpoint-epoch=*.ckpt"""), recursive=True)) _snake_case = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
655
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _snake_case = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _snake_case = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def _A ( __magic_name__ ): lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__magic_name__ )[0] @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) lowercase__ = data.reshape(__magic_name__ , __magic_name__ , __magic_name__ , 1 ) return data @deprecated(__magic_name__ , "Please use tf.one_hot on tensors." ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__magic_name__ ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=10 ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(__magic_name__ ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__magic_name__ , __magic_name__ ) return labels class lowerCAmelCase : @deprecated( _lowercase , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Tuple=False , _lowercase :str=False , _lowercase :Dict=dtypes.floataa , _lowercase :Optional[Any]=True , _lowercase :Any=None , ): '''simple docstring''' lowercase__ , lowercase__ = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowercase__ = 1_00_00 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'''images.shape: {images.shape} labels.shape: {labels.shape}''' lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(_lowercase , 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._images @property def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self._labels @property def UpperCAmelCase ( self :Dict ): '''simple docstring''' return self._num_examples @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._epochs_completed def UpperCAmelCase ( self :str , _lowercase :Union[str, Any] , _lowercase :Any=False , _lowercase :Union[str, Any]=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 7_84 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__magic_name__ , "Please write your own downloading logic." ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if not gfile.Exists(__magic_name__ ): gfile.MakeDirs(__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) if not gfile.Exists(__magic_name__ ): urllib.request.urlretrieve(__magic_name__ , __magic_name__ ) # noqa: S310 with gfile.GFile(__magic_name__ ) as f: lowercase__ = f.size() print("Successfully downloaded" , __magic_name__ , __magic_name__ , "bytes." ) return filepath @deprecated( __magic_name__ , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=False , __magic_name__=dtypes.floataa , __magic_name__=True , __magic_name__=5000 , __magic_name__=None , __magic_name__=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__magic_name__ , one_hot=__magic_name__ , dtype=__magic_name__ , seed=__magic_name__ ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = "train-images-idx3-ubyte.gz" lowercase__ = "train-labels-idx1-ubyte.gz" lowercase__ = "t10k-images-idx3-ubyte.gz" lowercase__ = "t10k-labels-idx1-ubyte.gz" lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) if not 0 <= validation_size <= len(__magic_name__ ): lowercase__ = ( "Validation size should be between 0 and " f'''{len(__magic_name__ )}. Received: {validation_size}.''' ) raise ValueError(__magic_name__ ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {"dtype": dtype, "reshape": reshape, "seed": seed} lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ )
655
1
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'AutoTokenizer' __lowerCamelCase = ['tokenizer'] __lowerCamelCase = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self :Dict , _lowercase :List[str] , _lowercase :List[Any]=None ): '''simple docstring''' super().__init__(_lowercase ) lowercase__ = speaker_embeddings @classmethod def UpperCAmelCase ( cls :Any , _lowercase :int , _lowercase :str="speaker_embeddings_path.json" , **_lowercase :List[str] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase__ = get_file_from_repo( _lowercase , _lowercase , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowercase , _lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase__ = None else: with open(_lowercase ) as speaker_embeddings_json: lowercase__ = json.load(_lowercase ) else: lowercase__ = None lowercase__ = AutoTokenizer.from_pretrained(_lowercase , **_lowercase ) return cls(tokenizer=_lowercase , speaker_embeddings=_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :List[str]="speaker_embeddings_path.json" , _lowercase :Any="speaker_embeddings" , _lowercase :bool = False , **_lowercase :Any , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowercase , _lowercase , "v2" ) , exist_ok=_lowercase ) lowercase__ = {} lowercase__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase__ = self._load_voice_preset(_lowercase ) lowercase__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowercase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowercase , ) lowercase__ = os.path.join(_lowercase , f'''{prompt_key}_{key}.npy''' ) lowercase__ = tmp_dict with open(os.path.join(_lowercase , _lowercase ) , "w" ) as fp: json.dump(_lowercase , _lowercase ) super().save_pretrained(_lowercase , _lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :str = None , **_lowercase :List[Any] ): '''simple docstring''' lowercase__ = self.speaker_embeddings[voice_preset] lowercase__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase__ = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) lowercase__ = np.load(_lowercase ) return voice_preset_dict def UpperCAmelCase ( self :Optional[int] , _lowercase :Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Optional[Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=None , _lowercase :List[str]="pt" , _lowercase :List[Any]=2_56 , _lowercase :List[str]=False , _lowercase :Union[str, Any]=True , _lowercase :Dict=False , **_lowercase :Tuple , ): '''simple docstring''' if voice_preset is not None and not isinstance(_lowercase , _lowercase ): if ( isinstance(_lowercase , _lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase__ = self._load_voice_preset(_lowercase ) else: if isinstance(_lowercase , _lowercase ) and not voice_preset.endswith(".npz" ): lowercase__ = voice_preset + ".npz" lowercase__ = np.load(_lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowercase , **_lowercase ) lowercase__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) lowercase__ = self.tokenizer( _lowercase , return_tensors=_lowercase , padding="max_length" , max_length=_lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , add_special_tokens=_lowercase , **_lowercase , ) if voice_preset is not None: lowercase__ = voice_preset return encoded_text
655
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
1
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class lowerCAmelCase ( ctypes.Structure ): # _fields is a specific attr expected by ctypes __lowerCamelCase = [('size', ctypes.c_int), ('visible', ctypes.c_byte)] def _A ( ): if os.name == "nt": lowercase__ = CursorInfo() lowercase__ = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__magic_name__ , ctypes.byref(__magic_name__ ) ) lowercase__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(__magic_name__ , ctypes.byref(__magic_name__ ) ) elif os.name == "posix": sys.stdout.write("\033[?25l" ) sys.stdout.flush() def _A ( ): if os.name == "nt": lowercase__ = CursorInfo() lowercase__ = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__magic_name__ , ctypes.byref(__magic_name__ ) ) lowercase__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(__magic_name__ , ctypes.byref(__magic_name__ ) ) elif os.name == "posix": sys.stdout.write("\033[?25h" ) sys.stdout.flush() @contextmanager def _A ( ): try: hide_cursor() yield finally: show_cursor()
655
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
1
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever _snake_case = logging.getLogger(__name__) class lowerCAmelCase ( lowercase_ ): def __init__( self :Optional[int] , _lowercase :Tuple , _lowercase :Tuple , _lowercase :Union[str, Any] , _lowercase :str=None ): '''simple docstring''' super().__init__( _lowercase , question_encoder_tokenizer=_lowercase , generator_tokenizer=_lowercase , index=_lowercase , init_retrieval=_lowercase , ) lowercase__ = None def UpperCAmelCase ( self :Dict , _lowercase :int ): '''simple docstring''' logger.info("initializing retrieval" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("dist initialized" ) # needs to be set manually lowercase__ = self._infer_socket_ifname() # avoid clash with the NCCL port lowercase__ = str(distributed_port + 1 ) lowercase__ = dist.new_group(ranks=_lowercase , backend="gloo" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("dist not initialized / main" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' return dist.get_rank(group=self.process_group ) == 0 def UpperCAmelCase ( self :int , _lowercase :Optional[Any] , _lowercase :Any , _lowercase :Union[str, Any]=torch.floataa ): '''simple docstring''' lowercase__ = torch.empty(_lowercase , dtype=_lowercase ) dist.scatter(_lowercase , src=0 , scatter_list=_lowercase , group=self.process_group ) return target_tensor def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = psutil.net_if_addrs() # a hacky way to deal with varying network interface names lowercase__ = next((addr for addr in addrs if addr.startswith("e" )) , _lowercase ) return ifname def UpperCAmelCase ( self :Optional[Any] , _lowercase :np.ndarray , _lowercase :int ): '''simple docstring''' if not dist.is_initialized(): lowercase__ , lowercase__ = self._main_retrieve(_lowercase , _lowercase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_lowercase ) # distributed training lowercase__ = dist.get_world_size(group=self.process_group ) # gather logic lowercase__ = None if self._is_main(): lowercase__ = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(_lowercase )] dist.gather(torch.tensor(_lowercase ) , dst=0 , gather_list=_lowercase , group=self.process_group ) # scatter logic lowercase__ = question_hidden_states.shape[0] lowercase__ = [] lowercase__ = [] if self._is_main(): assert len(_lowercase ) == world_size lowercase__ , lowercase__ = self._main_retrieve(torch.cat(_lowercase ).numpy() , _lowercase ) lowercase__ , lowercase__ = torch.tensor(_lowercase ), torch.tensor(_lowercase ) lowercase__ = self._chunk_tensor(_lowercase , _lowercase ) lowercase__ = self._chunk_tensor(_lowercase , _lowercase ) lowercase__ = self._scattered(_lowercase , [n_queries, n_docs] , target_type=torch.intaa ) lowercase__ = self._scattered(_lowercase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(_lowercase )
655
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["prompt"] lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] if "image" in inputs: lowercase__ = inputs["image"] else: lowercase__ = None if "mask_image" in inputs: lowercase__ = inputs["mask_image"] else: lowercase__ = None if "original_image" in inputs: lowercase__ = inputs["original_image"] else: lowercase__ = None lowercase__ , lowercase__ = pipe.encode_prompt(_lowercase ) # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_lowercase , _lowercase , _lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowercase , _lowercase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 )
655
1
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def _A ( __magic_name__="" ): lowercase__ = tempfile.mkdtemp() return os.path.join(__magic_name__ , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = torch.rand(12 , dtype=torch.floataa ) - 0.5 lowercase__ = AgentAudio(_lowercase ) lowercase__ = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(_lowercase , agent_type.to_raw() , atol=1e-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(_lowercase ) ) # Ensure that the file contains the same value as the original tensor lowercase__ , lowercase__ = sf.read(_lowercase ) self.assertTrue(torch.allclose(_lowercase , torch.tensor(_lowercase ) , atol=1e-4 ) ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = torch.rand(12 , dtype=torch.floataa ) - 0.5 lowercase__ = get_new_path(suffix=".wav" ) sf.write(_lowercase , _lowercase , 1_60_00 ) lowercase__ = AgentAudio(_lowercase ) self.assertTrue(torch.allclose(_lowercase , agent_type.to_raw() , atol=1e-4 ) ) self.assertEqual(agent_type.to_string() , _lowercase ) @require_vision @require_torch class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = torch.randint(0 , 2_56 , (64, 64, 3) ) lowercase__ = AgentImage(_lowercase ) lowercase__ = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(_lowercase , agent_type._tensor , atol=1e-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_lowercase ) ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" lowercase__ = Image.open(_lowercase ) lowercase__ = AgentImage(_lowercase ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_lowercase ) ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" lowercase__ = Image.open(_lowercase ) lowercase__ = AgentImage(_lowercase ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(_lowercase ) ) class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = "Hey!" lowercase__ = AgentText(_lowercase ) self.assertEqual(_lowercase , agent_type.to_string() ) self.assertEqual(_lowercase , agent_type.to_raw() ) self.assertEqual(_lowercase , _lowercase )
655
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) lowercase__ = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase__ = model(_lowercase )["last_hidden_state"] lowercase__ = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
655
1
def _A ( __magic_name__ , __magic_name__ ): return 1 if input_a == input_a else 0 def _A ( ): assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
655
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
1
from __future__ import annotations _snake_case = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ): lowercase__ = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the reference grid lowercase__ = 1 lowercase__ = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the action grid lowercase__ = init[0] lowercase__ = init[1] lowercase__ = 0 lowercase__ = g + heuristic[x][y] # cost from starting cell to destination cell lowercase__ = [[f, g, x, y]] lowercase__ = False # flag that is set when search is complete lowercase__ = False # flag set if we can't find expand while not found and not resign: if len(__magic_name__ ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() lowercase__ = cell.pop() lowercase__ = next_cell[2] lowercase__ = next_cell[3] lowercase__ = next_cell[1] if x == goal[0] and y == goal[1]: lowercase__ = True else: for i in range(len(__magic_name__ ) ): # to try out different valid actions lowercase__ = x + DIRECTIONS[i][0] lowercase__ = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__magic_name__ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: lowercase__ = g + cost lowercase__ = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) lowercase__ = 1 lowercase__ = i lowercase__ = [] lowercase__ = goal[0] lowercase__ = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: lowercase__ = x - DIRECTIONS[action[x][y]][0] lowercase__ = y - DIRECTIONS[action[x][y]][1] lowercase__ = xa lowercase__ = ya invpath.append([x, y] ) lowercase__ = [] for i in range(len(__magic_name__ ) ): path.append(invpath[len(__magic_name__ ) - 1 - i] ) return path, action if __name__ == "__main__": _snake_case = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] _snake_case = [0, 0] # all coordinates are given in format [y,x] _snake_case = [len(grid) - 1, len(grid[0]) - 1] _snake_case = 1 # the cost map which pushes the path closer to the goal _snake_case = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): _snake_case = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map _snake_case = 99 _snake_case , _snake_case = search(grid, init, goal, cost, heuristic) print("""ACTION MAP""") for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
655
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
1
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline _snake_case = argparse.ArgumentParser("""Stable Diffusion script with intel optimization""", add_help=False) parser.add_argument("""--dpm""", action="""store_true""", help="""Enable DPMSolver or not""") parser.add_argument("""--steps""", default=None, type=int, help="""Num inference steps""") _snake_case = parser.parse_args() _snake_case = """cpu""" _snake_case = """a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings""" _snake_case = """path-to-your-trained-model""" _snake_case = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: _snake_case = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) _snake_case = pipe.to(device) # to channels last _snake_case = pipe.unet.to(memory_format=torch.channels_last) _snake_case = pipe.vae.to(memory_format=torch.channels_last) _snake_case = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: _snake_case = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex _snake_case = torch.randn(2, 4, 64, 64) _snake_case = torch.rand(1) * 999 _snake_case = torch.randn(2, 77, 768) _snake_case = (sample, timestep, encoder_hidden_status) try: _snake_case = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: _snake_case = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) _snake_case = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) _snake_case = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: _snake_case = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute _snake_case = 666 _snake_case = torch.Generator(device).manual_seed(seed) _snake_case = {"""generator""": generator} if args.steps is not None: _snake_case = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): _snake_case = pipe(prompt, **generate_kwargs).images[0] # save image image.save("""generated.png""")
655
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = SMALL_MODEL_IDENTIFIER lowercase__ = "pt" lowercase__ = "tf" def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :int ): '''simple docstring''' lowercase__ = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = "mock_framework" # Framework provided - return whatever the user provides lowercase__ = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model )
655
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _snake_case = logging.get_logger(__name__) _snake_case = { """google/bit-50""": """https://huggingface.co/google/bit-50/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = 'bit' __lowerCamelCase = ['preactivation', 'bottleneck'] __lowerCamelCase = ['SAME', 'VALID'] def __init__( self :Union[str, Any] , _lowercase :List[str]=3 , _lowercase :Union[str, Any]=64 , _lowercase :List[str]=[2_56, 5_12, 10_24, 20_48] , _lowercase :Tuple=[3, 4, 6, 3] , _lowercase :Any="preactivation" , _lowercase :int="relu" , _lowercase :Dict=None , _lowercase :List[Any]=32 , _lowercase :Dict=0.0 , _lowercase :List[str]=False , _lowercase :Any=32 , _lowercase :List[Any]=1 , _lowercase :Union[str, Any]=None , _lowercase :List[str]=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(**_lowercase ) if layer_type not in self.layer_types: raise ValueError(f'''layer_type={layer_type} is not one of {','.join(self.layer_types )}''' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: lowercase__ = global_padding.upper() else: raise ValueError(f'''Padding strategy {global_padding} not supported''' ) lowercase__ = num_channels lowercase__ = embedding_size lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = layer_type lowercase__ = hidden_act lowercase__ = global_padding lowercase__ = num_groups lowercase__ = drop_path_rate lowercase__ = embedding_dynamic_padding lowercase__ = output_stride lowercase__ = width_factor lowercase__ = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(_lowercase ) + 1 )] lowercase__ , lowercase__ = get_aligned_output_features_output_indices( out_features=_lowercase , out_indices=_lowercase , stage_names=self.stage_names )
655
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git_vision_model' def __init__( self :Dict , _lowercase :Dict=7_68 , _lowercase :Dict=30_72 , _lowercase :Tuple=12 , _lowercase :List[str]=12 , _lowercase :Tuple=3 , _lowercase :Dict=2_24 , _lowercase :Tuple=16 , _lowercase :Optional[int]="quick_gelu" , _lowercase :Union[str, Any]=1e-5 , _lowercase :Tuple=0.0 , _lowercase :Tuple=0.02 , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = hidden_size lowercase__ = intermediate_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_channels lowercase__ = patch_size lowercase__ = image_size lowercase__ = initializer_range lowercase__ = attention_dropout lowercase__ = layer_norm_eps lowercase__ = hidden_act @classmethod def UpperCAmelCase ( cls :List[str] , _lowercase :Union[str, os.PathLike] , **_lowercase :Optional[int] ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git' def __init__( self :Union[str, Any] , _lowercase :Dict=None , _lowercase :List[str]=3_05_22 , _lowercase :Tuple=7_68 , _lowercase :Any=6 , _lowercase :Dict=12 , _lowercase :Any=30_72 , _lowercase :List[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :Optional[int]=0.1 , _lowercase :Optional[Any]=10_24 , _lowercase :Any=0.02 , _lowercase :int=1e-12 , _lowercase :List[Any]=0 , _lowercase :int="absolute" , _lowercase :List[str]=True , _lowercase :Any=False , _lowercase :int=1_01 , _lowercase :str=1_02 , _lowercase :Dict=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , pad_token_id=_lowercase , **_lowercase ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowercase__ = GitVisionConfig(**_lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = tie_word_embeddings lowercase__ = num_image_with_embedding lowercase__ = bos_token_id lowercase__ = eos_token_id def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
655
1
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 @add_end_docstrings(lowercase_ ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self :Any , *_lowercase :Optional[Any] , **_lowercase :Union[str, Any] ): '''simple docstring''' super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowercase__ = None if self.model.config.prefix is not None: lowercase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowercase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowercase__ , lowercase__ , lowercase__ = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) lowercase__ = {**self._preprocess_params, **preprocess_params} lowercase__ = {**self._forward_params, **forward_params} def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any]=None , _lowercase :List[Any]=None , _lowercase :List[str]=None , _lowercase :Optional[Any]=None , _lowercase :Optional[int]=None , _lowercase :Any=None , _lowercase :Any=None , _lowercase :Dict=None , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ = {} if prefix is not None: lowercase__ = prefix if prefix: lowercase__ = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) lowercase__ = handle_long_generation preprocess_params.update(_lowercase ) lowercase__ = generate_kwargs lowercase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.TENSORS if return_type is not None: lowercase__ = return_type if clean_up_tokenization_spaces is not None: lowercase__ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase ( self :int , *_lowercase :Optional[int] , **_lowercase :List[str] ): '''simple docstring''' if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self :Union[str, Any] , _lowercase :Dict , **_lowercase :Tuple ): '''simple docstring''' return super().__call__(_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple , _lowercase :Optional[int]="" , _lowercase :Tuple=None , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prompt_text if handle_long_generation == "hole": lowercase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: lowercase__ = generate_kwargs["max_new_tokens"] else: lowercase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: lowercase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) lowercase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: lowercase__ = inputs["attention_mask"][:, -keep_length:] return inputs def UpperCAmelCase ( self :str , _lowercase :int , **_lowercase :str ): '''simple docstring''' lowercase__ = model_inputs["input_ids"] lowercase__ = model_inputs.get("attention_mask" , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: lowercase__ = None lowercase__ = None lowercase__ = 1 else: lowercase__ = input_ids.shape[0] lowercase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowercase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: lowercase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: lowercase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowercase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowercase__ = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) lowercase__ = generated_sequence.shape[0] if self.framework == "pt": lowercase__ = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": lowercase__ = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCAmelCase ( self :Any , _lowercase :Tuple , _lowercase :str=ReturnType.FULL_TEXT , _lowercase :Dict=True ): '''simple docstring''' lowercase__ = model_outputs["generated_sequence"][0] lowercase__ = model_outputs["input_ids"] lowercase__ = model_outputs["prompt_text"] lowercase__ = generated_sequence.numpy().tolist() lowercase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowercase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowercase__ = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowercase__ = 0 else: lowercase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: lowercase__ = prompt_text + text[prompt_length:] else: lowercase__ = text[prompt_length:] lowercase__ = {"generated_text": all_text} records.append(_lowercase ) return records
655
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
1
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _snake_case = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _snake_case = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def _A ( __magic_name__ ): lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__magic_name__ )[0] @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) lowercase__ = data.reshape(__magic_name__ , __magic_name__ , __magic_name__ , 1 ) return data @deprecated(__magic_name__ , "Please use tf.one_hot on tensors." ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__magic_name__ ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=10 ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(__magic_name__ ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__magic_name__ , __magic_name__ ) return labels class lowerCAmelCase : @deprecated( _lowercase , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Tuple=False , _lowercase :str=False , _lowercase :Dict=dtypes.floataa , _lowercase :Optional[Any]=True , _lowercase :Any=None , ): '''simple docstring''' lowercase__ , lowercase__ = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowercase__ = 1_00_00 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'''images.shape: {images.shape} labels.shape: {labels.shape}''' lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(_lowercase , 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._images @property def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self._labels @property def UpperCAmelCase ( self :Dict ): '''simple docstring''' return self._num_examples @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._epochs_completed def UpperCAmelCase ( self :str , _lowercase :Union[str, Any] , _lowercase :Any=False , _lowercase :Union[str, Any]=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 7_84 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__magic_name__ , "Please write your own downloading logic." ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if not gfile.Exists(__magic_name__ ): gfile.MakeDirs(__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) if not gfile.Exists(__magic_name__ ): urllib.request.urlretrieve(__magic_name__ , __magic_name__ ) # noqa: S310 with gfile.GFile(__magic_name__ ) as f: lowercase__ = f.size() print("Successfully downloaded" , __magic_name__ , __magic_name__ , "bytes." ) return filepath @deprecated( __magic_name__ , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=False , __magic_name__=dtypes.floataa , __magic_name__=True , __magic_name__=5000 , __magic_name__=None , __magic_name__=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__magic_name__ , one_hot=__magic_name__ , dtype=__magic_name__ , seed=__magic_name__ ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = "train-images-idx3-ubyte.gz" lowercase__ = "train-labels-idx1-ubyte.gz" lowercase__ = "t10k-images-idx3-ubyte.gz" lowercase__ = "t10k-labels-idx1-ubyte.gz" lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) if not 0 <= validation_size <= len(__magic_name__ ): lowercase__ = ( "Validation size should be between 0 and " f'''{len(__magic_name__ )}. Received: {validation_size}.''' ) raise ValueError(__magic_name__ ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {"dtype": dtype, "reshape": reshape, "seed": seed} lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ )
655
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _snake_case = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val def _A ( __magic_name__ ): lowercase__ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowercase__ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) lowercase__ = value else: lowercase__ = value return new_state_dict def _A ( __magic_name__ ): lowercase__ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowercase__ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowercase__ = in_proj_weight_cross_attn[:256, :] lowercase__ = in_proj_bias_cross_attn[:256] lowercase__ = in_proj_weight_cross_attn[256:512, :] lowercase__ = in_proj_bias_cross_attn[256:512] lowercase__ = in_proj_weight_cross_attn[-256:, :] lowercase__ = in_proj_bias_cross_attn[-256:] def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = image.size lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 800 if "detection" in checkpoint_url else 1000 lowercase__ = target_max_size / current_max_size lowercase__ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _A ( __magic_name__ ): lowercase__ = F.to_tensor(__magic_name__ ) lowercase__ = F.normalize(__magic_name__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): logger.info("Converting model..." ) # load original state dict lowercase__ = torch.hub.load_state_dict_from_url(__magic_name__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowercase__ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val # create HuggingFace model and load state dict lowercase__ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: lowercase__ = 15 lowercase__ = 2 lowercase__ = {0: "table", 1: "table rotated"} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 125 lowercase__ = 6 lowercase__ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) lowercase__ = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion lowercase__ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" lowercase__ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=__magic_name__ ) lowercase__ = Image.open(__magic_name__ ).convert("RGB" ) lowercase__ = normalize(resize(__magic_name__ , __magic_name__ ) ).unsqueeze(0 ) lowercase__ = model(__magic_name__ ) if "detection" in checkpoint_url: lowercase__ = (1, 15, 3) lowercase__ = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) lowercase__ = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: lowercase__ = (1, 125, 7) lowercase__ = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) lowercase__ = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __magic_name__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) lowercase__ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _snake_case = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
655
1
from __future__ import annotations from functools import lru_cache from math import ceil _snake_case = 100 _snake_case = set(range(3, NUM_PRIMES, 2)) primes.add(2) _snake_case = 42 for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def _A ( __magic_name__ ): if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} lowercase__ = set() lowercase__ = 42 lowercase__ = 42 for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def _A ( __magic_name__ = 5000 ): for number_to_partition in range(1 , __magic_name__ ): if len(partition(__magic_name__ ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F"""{solution() = }""")
655
from typing import TYPE_CHECKING from ...utils import _LazyModule _snake_case = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
1
def _A ( __magic_name__ ): if not isinstance(__magic_name__ , __magic_name__ ): raise ValueError("Input must be an integer" ) if input_num <= 0: raise ValueError("Input must be positive" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
655
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = """ Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _A ( __magic_name__ , __magic_name__ , __magic_name__=8 ): lowercase__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :UNetaDConditionModel , _lowercase :DDPMScheduler , _lowercase :VQModel , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) lowercase__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :int , _lowercase :str ): '''simple docstring''' if latents is None: lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase__ = latents.to(_lowercase ) lowercase__ = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self :int , _lowercase :int=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) lowercase__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase__ = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase__ , lowercase__ = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. lowercase__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__( self :int , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 1_00 , _lowercase :float = 4.0 , _lowercase :int = 1 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = self._execution_device lowercase__ = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) lowercase__ = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: lowercase__ = image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) lowercase__ = self.scheduler.timesteps lowercase__ = self.unet.config.in_channels lowercase__ , lowercase__ = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) # create initial latent lowercase__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowercase , _lowercase , _lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance lowercase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase__ = {"image_embeds": image_embeds} lowercase__ = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) lowercase__ , lowercase__ = noise_pred.chunk(2 ) lowercase__ , lowercase__ = variance_pred.chunk(2 ) lowercase__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase__ = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing lowercase__ = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase__ = image * 0.5 + 0.5 lowercase__ = image.clamp(0 , 1 ) lowercase__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
655
1
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = StableUnCLIPImgaImgPipeline __lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS __lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __lowerCamelCase = frozenset([] ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = 32 lowercase__ = embedder_hidden_size # image encoding components lowercase__ = CLIPImageProcessor(crop_size=32 , size=32 ) torch.manual_seed(0 ) lowercase__ = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=_lowercase , projection_dim=_lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) lowercase__ = StableUnCLIPImageNormalizer(embedding_dim=_lowercase ) lowercase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) lowercase__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowercase__ = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_lowercase , layers_per_block=1 , upcast_attention=_lowercase , use_linear_projection=_lowercase , ) torch.manual_seed(0 ) lowercase__ = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.00085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=_lowercase , steps_offset=1 , ) torch.manual_seed(0 ) lowercase__ = AutoencoderKL() lowercase__ = { # image encoding components "feature_extractor": feature_extractor, "image_encoder": image_encoder.eval(), # image noising components "image_normalizer": image_normalizer.eval(), "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder.eval(), "unet": unet.eval(), "scheduler": scheduler, "vae": vae.eval(), } return components def UpperCAmelCase ( self :Dict , _lowercase :Optional[int] , _lowercase :int=0 , _lowercase :int=True ): '''simple docstring''' if str(_lowercase ).startswith("mps" ): lowercase__ = torch.manual_seed(_lowercase ) else: lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowercase ) ).to(_lowercase ) if pil_image: lowercase__ = input_image * 0.5 + 0.5 lowercase__ = input_image.clamp(0 , 1 ) lowercase__ = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() lowercase__ = DiffusionPipeline.numpy_to_pil(_lowercase )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = StableUnCLIPImgaImgPipeline(**_lowercase ) lowercase__ = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) inputs.update({"image_embeds": None} ) lowercase__ = sd_pipe(**_lowercase ).images lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase__ = np.array([0.3872, 0.7224, 0.5601, 0.4741, 0.6872, 0.5814, 0.4636, 0.3867, 0.5078] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = torch_device in ["cpu", "mps"] self._test_attention_slicing_forward_pass(test_max_difference=_lowercase ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=_lowercase ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_max_difference=_lowercase ) @slow @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :int ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) lowercase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy" ) lowercase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-l-img2img" , torch_dtype=torch.floataa ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowercase__ = torch.Generator(device="cpu" ).manual_seed(0 ) lowercase__ = pipe(_lowercase , "anime turle" , generator=_lowercase , output_type="np" ) lowercase__ = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(_lowercase , _lowercase ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) lowercase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy" ) lowercase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowercase__ = torch.Generator(device="cpu" ).manual_seed(0 ) lowercase__ = pipe(_lowercase , "anime turle" , generator=_lowercase , output_type="np" ) lowercase__ = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(_lowercase , _lowercase ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowercase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowercase__ = pipe( _lowercase , "anime turtle" , num_inference_steps=2 , output_type="np" , ) lowercase__ = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
655
import inspect import unittest class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :int ): '''simple docstring''' try: import diffusers # noqa: F401 except ImportError: assert False def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' import diffusers from diffusers.dependency_versions_table import deps lowercase__ = inspect.getmembers(_lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowercase__ = "k-diffusion" elif backend == "invisible_watermark": lowercase__ = "invisible-watermark" assert backend in deps, f'''{backend} is not in the deps table!'''
655
1
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = 1.5 lowercase__ = int(factor * num_class_images ) lowercase__ = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=__magic_name__ , aesthetic_weight=0.1 ) os.makedirs(f'''{class_data_dir}/images''' , exist_ok=__magic_name__ ) if len(list(Path(f'''{class_data_dir}/images''' ).iterdir() ) ) >= num_class_images: return while True: lowercase__ = client.query(text=__magic_name__ ) if len(__magic_name__ ) >= factor * num_class_images or num_images > 1e4: break else: lowercase__ = int(factor * num_images ) lowercase__ = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=__magic_name__ , aesthetic_weight=0.1 , ) lowercase__ = 0 lowercase__ = 0 lowercase__ = tqdm(desc="downloading real regularization images" , total=__magic_name__ ) with open(f'''{class_data_dir}/caption.txt''' , "w" ) as fa, open(f'''{class_data_dir}/urls.txt''' , "w" ) as fa, open( f'''{class_data_dir}/images.txt''' , "w" ) as fa: while total < num_class_images: lowercase__ = class_images[count] count += 1 try: lowercase__ = requests.get(images["url"] ) if img.status_code == 200: lowercase__ = Image.open(BytesIO(img.content ) ) with open(f'''{class_data_dir}/images/{total}.jpg''' , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f'''{class_data_dir}/images/{total}.jpg''' + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _A ( ): lowercase__ = argparse.ArgumentParser("" , add_help=__magic_name__ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=__magic_name__ , type=__magic_name__ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=__magic_name__ , type=__magic_name__ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=__magic_name__ ) return parser.parse_args() if __name__ == "__main__": _snake_case = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
655
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 # setable values __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None @classmethod def UpperCAmelCase ( cls :Union[str, Any] , _lowercase :CommonSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray ): '''simple docstring''' return cls(common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase ) @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCamelCase = 42 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return True @register_to_config def __init__( self :str , _lowercase :int = 10_00 , _lowercase :float = 0.0001 , _lowercase :float = 0.02 , _lowercase :str = "linear" , _lowercase :Optional[jnp.ndarray] = None , _lowercase :str = "fixed_small" , _lowercase :bool = True , _lowercase :str = "epsilon" , _lowercase :jnp.dtype = jnp.floataa , ): '''simple docstring''' lowercase__ = dtype def UpperCAmelCase ( self :str , _lowercase :Optional[CommonSchedulerState] = None ): '''simple docstring''' if common is None: lowercase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowercase__ = jnp.array(1.0 , dtype=self.dtype ) lowercase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase ( self :List[str] , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :Tuple = () ): '''simple docstring''' lowercase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowercase__ = (jnp.arange(0 , _lowercase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Tuple , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :List[str]=None , _lowercase :Tuple=None ): '''simple docstring''' lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowercase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowercase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowercase__ = jnp.clip(_lowercase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowercase__ = jnp.log(jnp.clip(_lowercase , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowercase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowercase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowercase__ = variance lowercase__ = state.common.betas[t] lowercase__ = (predicted_variance + 1) / 2 lowercase__ = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase ( self :Optional[int] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :int , _lowercase :jnp.ndarray , _lowercase :Optional[jax.random.KeyArray] = None , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = timestep if key is None: lowercase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ = jnp.split(_lowercase , sample.shape[1] , axis=1 ) else: lowercase__ = None # 1. compute alphas, betas lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowercase__ = 1 - alpha_prod_t lowercase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowercase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowercase__ = model_output elif self.config.prediction_type == "v_prediction": lowercase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowercase__ = jnp.clip(_lowercase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowercase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowercase__ = jax.random.split(_lowercase , num=1 ) lowercase__ = jax.random.normal(_lowercase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowercase , _lowercase , predicted_variance=_lowercase ) ** 0.5) * noise lowercase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowercase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowercase , state=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return add_noise_common(state.common , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return get_velocity_common(state.common , _lowercase , _lowercase , _lowercase ) def __len__( self :List[str] ): '''simple docstring''' return self.config.num_train_timesteps
655
1
def _A ( __magic_name__ = 200 ): lowercase__ = [1, 2, 5, 10, 20, 50, 100, 200] lowercase__ = [0] * (pence + 1) lowercase__ = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(__magic_name__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
655
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _snake_case = logging.get_logger(__name__) _snake_case = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _snake_case = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCAmelCase : __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Model type selected in the list: ' + ', '.join(lowercase_ )} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) __lowerCamelCase = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) __lowerCamelCase = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) __lowerCamelCase = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) __lowerCamelCase = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'train' __lowerCamelCase = 'dev' class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 def __init__( self :Optional[Any] , _lowercase :SquadDataTrainingArguments , _lowercase :PreTrainedTokenizer , _lowercase :Optional[int] = None , _lowercase :Union[str, Split] = Split.train , _lowercase :Optional[bool] = False , _lowercase :Optional[str] = None , _lowercase :Optional[str] = "pt" , ): '''simple docstring''' lowercase__ = args lowercase__ = is_language_sensitive lowercase__ = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_lowercase , _lowercase ): try: lowercase__ = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) lowercase__ = mode # Load data features from cache or dataset file lowercase__ = "v2" if args.version_2_with_negative else "v1" lowercase__ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase__ = cached_features_file + ".lock" with FileLock(_lowercase ): if os.path.exists(_lowercase ) and not args.overwrite_cache: lowercase__ = time.time() lowercase__ = torch.load(_lowercase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowercase__ = self.old_features["features"] lowercase__ = self.old_features.get("dataset" , _lowercase ) lowercase__ = self.old_features.get("examples" , _lowercase ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: lowercase__ = self.processor.get_dev_examples(args.data_dir ) else: lowercase__ = self.processor.get_train_examples(args.data_dir ) lowercase__ , lowercase__ = squad_convert_examples_to_features( examples=self.examples , tokenizer=_lowercase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_lowercase , ) lowercase__ = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _lowercase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self :Dict ): '''simple docstring''' return len(self.features ) def __getitem__( self :Any , _lowercase :Any ): '''simple docstring''' lowercase__ = self.features[i] lowercase__ = torch.tensor(feature.input_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.attention_mask , dtype=torch.long ) lowercase__ = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.cls_index , dtype=torch.long ) lowercase__ = torch.tensor(feature.p_mask , dtype=torch.float ) lowercase__ = torch.tensor(feature.is_impossible , dtype=torch.float ) lowercase__ = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowercase__ = torch.tensor(feature.start_position , dtype=torch.long ) lowercase__ = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
655
1
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ['image_processor', 'tokenizer'] __lowerCamelCase = 'BlipImageProcessor' __lowerCamelCase = 'AutoTokenizer' def __init__( self :Optional[Any] , _lowercase :Tuple , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = False super().__init__(_lowercase , _lowercase ) lowercase__ = self.image_processor def __call__( self :Union[str, Any] , _lowercase :ImageInput = None , _lowercase :Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , _lowercase :bool = True , _lowercase :Union[bool, str, PaddingStrategy] = False , _lowercase :Union[bool, str, TruncationStrategy] = None , _lowercase :Optional[int] = None , _lowercase :int = 0 , _lowercase :Optional[int] = None , _lowercase :Optional[bool] = None , _lowercase :bool = False , _lowercase :bool = False , _lowercase :bool = False , _lowercase :bool = False , _lowercase :bool = False , _lowercase :bool = True , _lowercase :Optional[Union[str, TensorType]] = None , **_lowercase :int , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: lowercase__ = self.tokenizer lowercase__ = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) return text_encoding # add pixel_values lowercase__ = self.image_processor(_lowercase , return_tensors=_lowercase ) if text is not None: lowercase__ = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) else: lowercase__ = None if text_encoding is not None: encoding_image_processor.update(_lowercase ) return encoding_image_processor def UpperCAmelCase ( self :List[Any] , *_lowercase :Any , **_lowercase :Optional[Any] ): '''simple docstring''' return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , *_lowercase :Dict , **_lowercase :Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*_lowercase , **_lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer.model_input_names lowercase__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
655
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = """▁""" _snake_case = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} _snake_case = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } _snake_case = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } _snake_case = { """ernie-m-base""": 514, """ernie-m-large""": 514, } _snake_case = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ["input_ids"] __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = RESOURCE_FILES_NAMES def __init__( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=False , _lowercase :Dict="utf8" , _lowercase :Optional[Any]="[UNK]" , _lowercase :Optional[int]="[SEP]" , _lowercase :List[str]="[PAD]" , _lowercase :Dict="[CLS]" , _lowercase :Optional[Any]="[MASK]" , _lowercase :Optional[Dict[str, Any]] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , vocab_file=_lowercase , encoding=_lowercase , sp_model_kwargs=self.sp_model_kwargs , **_lowercase , ) lowercase__ = do_lower_case lowercase__ = sentencepiece_model_ckpt lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase__ = self.load_vocab(filepath=_lowercase ) else: lowercase__ = {self.sp_model.id_to_piece(_lowercase ): id for id in range(self.sp_model.get_piece_size() )} lowercase__ = {v: k for k, v in self.vocab.items()} def UpperCAmelCase ( self :Any , _lowercase :Dict ): '''simple docstring''' if text is None: return None lowercase__ = self.tokenize(_lowercase ) lowercase__ , lowercase__ = "", [] for i, ch in enumerate(_lowercase ): if ch in self.SP_CHAR_MAPPING: lowercase__ = self.SP_CHAR_MAPPING.get(_lowercase ) else: lowercase__ = unicodedata.normalize("NFKC" , _lowercase ) if self.is_whitespace(_lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(_lowercase ) ) lowercase__ , lowercase__ , lowercase__ = normalized_text, [], 0 if self.do_lower_case: lowercase__ = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase__ = token[1:] lowercase__ = text[offset:].index(_lowercase ) + offset lowercase__ = start + len(_lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase__ = end return token_mapping @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return len(self.vocab ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self :Any ): '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self :Optional[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCAmelCase ( self :List[str] , _lowercase :Optional[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(_lowercase , _lowercase ) for c in text) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any]=False , _lowercase :Optional[int]=64 , _lowercase :Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get("enable_sampling" ) is True: lowercase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: lowercase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: lowercase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: lowercase__ = self.sp_model.EncodeAsPieces(_lowercase ) else: lowercase__ = self.sp_model.SampleEncodeAsPieces(_lowercase , _lowercase , _lowercase ) lowercase__ = [] for pi, piece in enumerate(_lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(_lowercase ) and pi != 0: new_pieces.append(_lowercase ) continue else: continue lowercase__ = 0 for i, chunk in enumerate(_lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(_lowercase ) or self.is_punct(_lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(_lowercase ) lowercase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i if len(_lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCAmelCase ( self :Tuple , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = self.convert_ids_to_tokens(_lowercase ) lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' return self.vocab.get(_lowercase , self.vocab.get(self.unk_token ) ) def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' return self.reverse_vocab.get(_lowercase , self.unk_token ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :Tuple=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCAmelCase ( self :Dict , _lowercase :int , _lowercase :Union[str, Any]=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Dict=None , _lowercase :Optional[Any]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(_lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(_lowercase ) + 1) + [1] * (len(_lowercase ) + 3) def UpperCAmelCase ( self :str , _lowercase :Optional[int] ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCAmelCase ( self :int , _lowercase :Dict ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCAmelCase ( self :List[str] , _lowercase :List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(_lowercase ) == 1: lowercase__ = unicodedata.category(_lowercase ) if cat == "Zs": return True return False def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = {} with io.open(_lowercase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(_lowercase ): lowercase__ = line.rstrip("\n" ) lowercase__ = int(_lowercase ) return token_to_idx def UpperCAmelCase ( self :List[str] , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' lowercase__ = 0 if os.path.isdir(_lowercase ): lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowercase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(_lowercase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowercase__ = token_index writer.write(token + "\n" ) index += 1 lowercase__ = os.path.join(_lowercase , "sentencepiece.bpe.model" ) with open(_lowercase , "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(_lowercase ) return (vocab_file,)
655
1
import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration _snake_case = pytest.mark.integration _snake_case = {"""comet"""} _snake_case = importlib.util.find_spec("""fairseq""") is not None _snake_case = {"""code_eval"""} _snake_case = os.name == """nt""" _snake_case = {"""bertscore""", """frugalscore""", """perplexity"""} _snake_case = importlib.util.find_spec("""transformers""") is not None def _A ( __magic_name__ ): @wraps(__magic_name__ ) def wrapper(self , __magic_name__ ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("\"test requires Fairseq\"" ) else: test_case(self , __magic_name__ ) return wrapper def _A ( __magic_name__ ): @wraps(__magic_name__ ) def wrapper(self , __magic_name__ ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("\"test requires transformers\"" ) else: test_case(self , __magic_name__ ) return wrapper def _A ( __magic_name__ ): @wraps(__magic_name__ ) def wrapper(self , __magic_name__ ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("\"test not supported on Windows\"" ) else: test_case(self , __magic_name__ ) return wrapper def _A ( ): lowercase__ = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("./metrics/*/" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( lowercase_ , lowercase_ , lowercase_ ) @local class lowerCAmelCase ( parameterized.TestCase ): __lowerCamelCase = {} __lowerCamelCase = None @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning" ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "[...]" lowercase__ = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , _lowercase ) ).module_path ) lowercase__ = datasets.load.import_main_class(metric_module.__name__ , dataset=_lowercase ) # check parameters lowercase__ = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(_lowercase , metric_module.__name__ ): with self.use_local_metrics(): try: lowercase__ = doctest.testmod(_lowercase , verbose=_lowercase , raise_on_error=_lowercase ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def UpperCAmelCase ( self :str , _lowercase :int ): '''simple docstring''' lowercase__ = "[...]" lowercase__ = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , _lowercase ) ).module_path ) # run doctest with self.use_local_metrics(): lowercase__ = doctest.testmod(_lowercase , verbose=_lowercase , raise_on_error=_lowercase ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def UpperCAmelCase ( self :Optional[Any] , _lowercase :Tuple , _lowercase :Optional[Any] ): '''simple docstring''' if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](_lowercase ): yield else: yield @contextmanager def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' def load_local_metric(_lowercase :Optional[int] , *_lowercase :Dict , **_lowercase :int ): return load_metric(os.path.join("metrics" , _lowercase ) , *_lowercase , **_lowercase ) with patch("datasets.load_metric" ) as mock_load_metric: lowercase__ = load_local_metric yield @classmethod def UpperCAmelCase ( cls :Optional[int] , _lowercase :List[str] ): '''simple docstring''' def wrapper(_lowercase :List[Any] ): lowercase__ = contextmanager(_lowercase ) lowercase__ = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("bleurt" ) def _A ( __magic_name__ ): import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("sv" , "" , "" ) # handle pytest cli flags class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :List[str] , _lowercase :Optional[int] ): '''simple docstring''' assert len(input_dict["input_ids"] ) == 2 return np.array([1.03, 1.04] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("bleurt.score._create_predictor" ) as mock_create_predictor: lowercase__ = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("bertscore" ) def _A ( __magic_name__ ): import torch def bert_cos_score_idf(__magic_name__ , __magic_name__ , *__magic_name__ , **__magic_name__ ): return torch.tensor([[1.0, 1.0, 1.0]] * len(__magic_name__ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("bert_score.scorer.get_model" ), patch( "bert_score.scorer.bert_cos_score_idf" ) as mock_bert_cos_score_idf: lowercase__ = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("comet" ) def _A ( __magic_name__ ): def load_from_checkpoint(__magic_name__ ): class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] , _lowercase :List[str] , *_lowercase :List[str] , **_lowercase :str ): '''simple docstring''' assert len(_lowercase ) == 2 lowercase__ = [0.19, 0.92] return scores, sum(_lowercase ) / len(_lowercase ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("comet.download_model" ) as mock_download_model: lowercase__ = None with patch("comet.load_from_checkpoint" ) as mock_load_from_checkpoint: lowercase__ = load_from_checkpoint yield def _A ( ): lowercase__ = load_metric(os.path.join("metrics" , "seqeval" ) ) lowercase__ = "ERROR" lowercase__ = f'''Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}''' with pytest.raises(__magic_name__ , match=re.escape(__magic_name__ ) ): metric.compute(predictions=[] , references=[] , scheme=__magic_name__ )
655
def _A ( __magic_name__ ): lowercase__ = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def _A ( __magic_name__ = 100 ): lowercase__ = 1 lowercase__ = 2 for i in range(2 , max_n + 1 ): lowercase__ = pre_numerator lowercase__ = 2 * i // 3 if i % 3 == 0 else 1 lowercase__ = cur_numerator lowercase__ = e_cont * pre_numerator + temp return sum_digits(__magic_name__ ) if __name__ == "__main__": print(F"""{solution() = }""")
655
1
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _snake_case = random.Random() def _A ( __magic_name__ , __magic_name__=1.0 , __magic_name__=None , __magic_name__=None ): if rng is None: lowercase__ = global_rng lowercase__ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCAmelCase ( unittest.TestCase ): def __init__( self :Union[str, Any] , _lowercase :str , _lowercase :List[Any]=7 , _lowercase :int=4_00 , _lowercase :Optional[int]=20_00 , _lowercase :str=10 , _lowercase :Tuple=1_60 , _lowercase :str=8 , _lowercase :List[str]=0.0 , _lowercase :Optional[Any]=40_00 , _lowercase :List[Any]=False , _lowercase :int=True , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = min_seq_length lowercase__ = max_seq_length lowercase__ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowercase__ = padding_value lowercase__ = sampling_rate lowercase__ = return_attention_mask lowercase__ = do_normalize lowercase__ = feature_size lowercase__ = chunk_length lowercase__ = hop_length def UpperCAmelCase ( self :Any ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCAmelCase ( self :Optional[Any] , _lowercase :Optional[Any]=False , _lowercase :List[str]=False ): '''simple docstring''' def _flatten(_lowercase :List[Any] ): return list(itertools.chain(*_lowercase ) ) if equal_length: lowercase__ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowercase__ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowercase__ = [np.asarray(_lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = WhisperFeatureExtractionTester(self ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ = feat_extract_first.save_pretrained(_lowercase )[0] check_json_file_has_correct_format(_lowercase ) lowercase__ = self.feature_extraction_class.from_pretrained(_lowercase ) lowercase__ = feat_extract_first.to_dict() lowercase__ = feat_extract_second.to_dict() lowercase__ = feat_extract_first.mel_filters lowercase__ = feat_extract_second.mel_filters self.assertTrue(np.allclose(_lowercase , _lowercase ) ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ = os.path.join(_lowercase , "feat_extract.json" ) feat_extract_first.to_json_file(_lowercase ) lowercase__ = self.feature_extraction_class.from_json_file(_lowercase ) lowercase__ = feat_extract_first.to_dict() lowercase__ = feat_extract_second.to_dict() lowercase__ = feat_extract_first.mel_filters lowercase__ = feat_extract_second.mel_filters self.assertTrue(np.allclose(_lowercase , _lowercase ) ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowercase__ = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] lowercase__ = [np.asarray(_lowercase ) for speech_input in speech_inputs] # Test feature size lowercase__ = feature_extractor(_lowercase , padding="max_length" , return_tensors="np" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input lowercase__ = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_features lowercase__ = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_features self.assertTrue(np.allclose(_lowercase , _lowercase , atol=1e-3 ) ) # Test batched lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_lowercase , _lowercase ): self.assertTrue(np.allclose(_lowercase , _lowercase , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. lowercase__ = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] lowercase__ = np.asarray(_lowercase ) lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_lowercase , _lowercase ): self.assertTrue(np.allclose(_lowercase , _lowercase , atol=1e-3 ) ) # Test truncation required lowercase__ = [floats_list((1, x) )[0] for x in range(2_00 , (feature_extractor.n_samples + 5_00) , 2_00 )] lowercase__ = [np.asarray(_lowercase ) for speech_input in speech_inputs] lowercase__ = [x[: feature_extractor.n_samples] for x in speech_inputs] lowercase__ = [np.asarray(_lowercase ) for speech_input in speech_inputs_truncated] lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features lowercase__ = feature_extractor(_lowercase , return_tensors="np" ).input_features for enc_seq_a, enc_seq_a in zip(_lowercase , _lowercase ): self.assertTrue(np.allclose(_lowercase , _lowercase , atol=1e-3 ) ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' import torch lowercase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowercase__ = np.random.rand(1_00 , 32 ).astype(np.floataa ) lowercase__ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowercase__ = feature_extractor.pad([{"input_features": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowercase__ = feature_extractor.pad([{"input_features": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :List[Any] ): '''simple docstring''' lowercase__ = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech lowercase__ = ds.sort("id" ).select(range(_lowercase ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on lowercase__ = self._load_datasamples(1 ) lowercase__ = WhisperFeatureExtractor() lowercase__ = feature_extractor(_lowercase , return_tensors="pt" ).input_features self.assertEqual(input_features.shape , (1, 80, 30_00) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , _lowercase , atol=1e-4 ) ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowercase__ = self._load_datasamples(1 )[0] lowercase__ = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_55_35 # Rescale to [0, 65535] to show issue lowercase__ = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=_lowercase )[0] self.assertTrue(np.all(np.mean(_lowercase ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(_lowercase ) - 1 ) < 1e-3 ) )
655
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'AutoTokenizer' __lowerCamelCase = ['tokenizer'] __lowerCamelCase = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self :Dict , _lowercase :List[str] , _lowercase :List[Any]=None ): '''simple docstring''' super().__init__(_lowercase ) lowercase__ = speaker_embeddings @classmethod def UpperCAmelCase ( cls :Any , _lowercase :int , _lowercase :str="speaker_embeddings_path.json" , **_lowercase :List[str] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase__ = get_file_from_repo( _lowercase , _lowercase , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowercase , _lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase__ = None else: with open(_lowercase ) as speaker_embeddings_json: lowercase__ = json.load(_lowercase ) else: lowercase__ = None lowercase__ = AutoTokenizer.from_pretrained(_lowercase , **_lowercase ) return cls(tokenizer=_lowercase , speaker_embeddings=_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :List[str]="speaker_embeddings_path.json" , _lowercase :Any="speaker_embeddings" , _lowercase :bool = False , **_lowercase :Any , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowercase , _lowercase , "v2" ) , exist_ok=_lowercase ) lowercase__ = {} lowercase__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase__ = self._load_voice_preset(_lowercase ) lowercase__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowercase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowercase , ) lowercase__ = os.path.join(_lowercase , f'''{prompt_key}_{key}.npy''' ) lowercase__ = tmp_dict with open(os.path.join(_lowercase , _lowercase ) , "w" ) as fp: json.dump(_lowercase , _lowercase ) super().save_pretrained(_lowercase , _lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :str = None , **_lowercase :List[Any] ): '''simple docstring''' lowercase__ = self.speaker_embeddings[voice_preset] lowercase__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase__ = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) lowercase__ = np.load(_lowercase ) return voice_preset_dict def UpperCAmelCase ( self :Optional[int] , _lowercase :Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Optional[Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=None , _lowercase :List[str]="pt" , _lowercase :List[Any]=2_56 , _lowercase :List[str]=False , _lowercase :Union[str, Any]=True , _lowercase :Dict=False , **_lowercase :Tuple , ): '''simple docstring''' if voice_preset is not None and not isinstance(_lowercase , _lowercase ): if ( isinstance(_lowercase , _lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase__ = self._load_voice_preset(_lowercase ) else: if isinstance(_lowercase , _lowercase ) and not voice_preset.endswith(".npz" ): lowercase__ = voice_preset + ".npz" lowercase__ = np.load(_lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowercase , **_lowercase ) lowercase__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) lowercase__ = self.tokenizer( _lowercase , return_tensors=_lowercase , padding="max_length" , max_length=_lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , add_special_tokens=_lowercase , **_lowercase , ) if voice_preset is not None: lowercase__ = voice_preset return encoded_text
655
1
import os import sys import unittest _snake_case = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path _snake_case = os.path.join(git_repo_path, """src""", """transformers""") _snake_case = """ {0} = None """ _snake_case = """ class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) """ _snake_case = """ def {0}(*args, **kwargs): requires_backends({0}, {1}) """ class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(_lowercase ) lowercase__ = find_backend(" if not is_tokenizers_available():" ) self.assertEqual(_lowercase , "tokenizers" ) lowercase__ = find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(_lowercase , "tensorflow_text" ) lowercase__ = find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(_lowercase , "sentencepiece_and_tokenizers" ) lowercase__ = find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(_lowercase , "sentencepiece_and_tensorflow_text" ) lowercase__ = find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(_lowercase , "sentencepiece_and_tokenizers_and_vision" ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch" , _lowercase ) self.assertIn("tensorflow_text" , _lowercase ) self.assertIn("sentencepiece_and_tokenizers" , _lowercase ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertModel" , objects["tf"] ) self.assertIn("FlaxBertModel" , objects["flax"] ) self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertTokenizer" , objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer" , objects["sentencepiece_and_tokenizers"] ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = create_dummy_object("CONSTANT" , "'torch'" ) self.assertEqual(_lowercase , "\nCONSTANT = None\n" ) lowercase__ = create_dummy_object("function" , "'torch'" ) self.assertEqual( _lowercase , "\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n" ) lowercase__ = "\nclass FakeClass(metaclass=DummyObject):\n _backends = 'torch'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, 'torch')\n" lowercase__ = create_dummy_object("FakeClass" , "'torch'" ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" lowercase__ = create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"] , _lowercase )
655
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__magic_name__ ): # Forward propagation lowercase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? lowercase__ = (expected / 100) - layer_a # Error delta lowercase__ = layer_1_error * sigmoid_function(__magic_name__ , __magic_name__ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input("""Expected value: """)) _snake_case = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
655
1
def _A ( __magic_name__ ): if not isinstance(__magic_name__ , __magic_name__ ): raise ValueError("Input series is not valid, valid series - [2, 4, 6]" ) if len(__magic_name__ ) == 0: raise ValueError("Input list must be a non empty list" ) if len(__magic_name__ ) == 1: return True lowercase__ = series[1] - series[0] for index in range(len(__magic_name__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _A ( __magic_name__ ): if not isinstance(__magic_name__ , __magic_name__ ): raise ValueError("Input series is not valid, valid series - [2, 4, 6]" ) if len(__magic_name__ ) == 0: raise ValueError("Input list must be a non empty list" ) lowercase__ = 0 for val in series: answer += val return answer / len(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'van' def __init__( self :Optional[Any] , _lowercase :Dict=2_24 , _lowercase :Union[str, Any]=3 , _lowercase :List[Any]=[7, 3, 3, 3] , _lowercase :Any=[4, 2, 2, 2] , _lowercase :Union[str, Any]=[64, 1_28, 3_20, 5_12] , _lowercase :List[Any]=[3, 3, 12, 3] , _lowercase :Dict=[8, 8, 4, 4] , _lowercase :int="gelu" , _lowercase :List[Any]=0.02 , _lowercase :List[Any]=1e-6 , _lowercase :Any=1e-2 , _lowercase :int=0.0 , _lowercase :int=0.0 , **_lowercase :Dict , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = image_size lowercase__ = num_channels lowercase__ = patch_sizes lowercase__ = strides lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = mlp_ratios lowercase__ = hidden_act lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = layer_scale_init_value lowercase__ = drop_path_rate lowercase__ = dropout_rate
655
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def _A ( __magic_name__ ): lowercase__ = SwinvaConfig() lowercase__ = swinva_name.split("_" ) lowercase__ = name_split[1] if "to" in name_split[3]: lowercase__ = int(name_split[3][-3:] ) else: lowercase__ = int(name_split[3] ) if "to" in name_split[2]: lowercase__ = int(name_split[2][-2:] ) else: lowercase__ = int(name_split[2][6:] ) if model_size == "tiny": lowercase__ = 96 lowercase__ = (2, 2, 6, 2) lowercase__ = (3, 6, 12, 24) elif model_size == "small": lowercase__ = 96 lowercase__ = (2, 2, 18, 2) lowercase__ = (3, 6, 12, 24) elif model_size == "base": lowercase__ = 128 lowercase__ = (2, 2, 18, 2) lowercase__ = (4, 8, 16, 32) else: lowercase__ = 192 lowercase__ = (2, 2, 18, 2) lowercase__ = (6, 12, 24, 48) if "to" in swinva_name: lowercase__ = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): lowercase__ = 2_1841 lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-22k-id2label.json" lowercase__ = json.load(open(hf_hub_download(__magic_name__ , __magic_name__ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(__magic_name__ ): v for k, v in idalabel.items()} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 1000 lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(__magic_name__ , __magic_name__ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(__magic_name__ ): v for k, v in idalabel.items()} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = img_size lowercase__ = num_classes lowercase__ = embed_dim lowercase__ = depths lowercase__ = num_heads lowercase__ = window_size return config def _A ( __magic_name__ ): if "patch_embed.proj" in name: lowercase__ = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: lowercase__ = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: lowercase__ = "encoder." + name if "attn.proj" in name: lowercase__ = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: lowercase__ = name.replace("attn" , "attention.self" ) if "norm1" in name: lowercase__ = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: lowercase__ = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: lowercase__ = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: lowercase__ = name.replace("mlp.fc2" , "output.dense" ) if "q_bias" in name: lowercase__ = name.replace("q_bias" , "query.bias" ) if "k_bias" in name: lowercase__ = name.replace("k_bias" , "key.bias" ) if "v_bias" in name: lowercase__ = name.replace("v_bias" , "value.bias" ) if "cpb_mlp" in name: lowercase__ = name.replace("cpb_mlp" , "continuous_position_bias_mlp" ) if name == "norm.weight": lowercase__ = "layernorm.weight" if name == "norm.bias": lowercase__ = "layernorm.bias" if "head" in name: lowercase__ = name.replace("head" , "classifier" ) else: lowercase__ = "swinv2." + name return name def _A ( __magic_name__ , __magic_name__ ): for key in orig_state_dict.copy().keys(): lowercase__ = orig_state_dict.pop(__magic_name__ ) if "mask" in key: continue elif "qkv" in key: lowercase__ = key.split("." ) lowercase__ = int(key_split[1] ) lowercase__ = int(key_split[3] ) lowercase__ = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowercase__ = val[:dim, :] lowercase__ = val[dim : dim * 2, :] lowercase__ = val[-dim:, :] else: lowercase__ = val[:dim] lowercase__ = val[ dim : dim * 2 ] lowercase__ = val[-dim:] else: lowercase__ = val return orig_state_dict def _A ( __magic_name__ , __magic_name__ ): lowercase__ = timm.create_model(__magic_name__ , pretrained=__magic_name__ ) timm_model.eval() lowercase__ = get_swinva_config(__magic_name__ ) lowercase__ = SwinvaForImageClassification(__magic_name__ ) model.eval() lowercase__ = convert_state_dict(timm_model.state_dict() , __magic_name__ ) model.load_state_dict(__magic_name__ ) lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = AutoImageProcessor.from_pretrained("microsoft/{}".format(swinva_name.replace("_" , "-" ) ) ) lowercase__ = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) lowercase__ = image_processor(images=__magic_name__ , return_tensors="pt" ) lowercase__ = timm_model(inputs["pixel_values"] ) lowercase__ = model(**__magic_name__ ).logits assert torch.allclose(__magic_name__ , __magic_name__ , atol=1e-3 ) print(f'''Saving model {swinva_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__magic_name__ ) model.push_to_hub( repo_path_or_name=Path(__magic_name__ , __magic_name__ ) , organization="nandwalritik" , commit_message="Add model" , ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swinv2_name""", default="""swinv2_tiny_patch4_window8_256""", type=str, help="""Name of the Swinv2 timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _snake_case = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
655
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 @add_end_docstrings(lowercase_ ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self :Any , *_lowercase :Optional[Any] , **_lowercase :Union[str, Any] ): '''simple docstring''' super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowercase__ = None if self.model.config.prefix is not None: lowercase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowercase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowercase__ , lowercase__ , lowercase__ = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) lowercase__ = {**self._preprocess_params, **preprocess_params} lowercase__ = {**self._forward_params, **forward_params} def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any]=None , _lowercase :List[Any]=None , _lowercase :List[str]=None , _lowercase :Optional[Any]=None , _lowercase :Optional[int]=None , _lowercase :Any=None , _lowercase :Any=None , _lowercase :Dict=None , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ = {} if prefix is not None: lowercase__ = prefix if prefix: lowercase__ = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) lowercase__ = handle_long_generation preprocess_params.update(_lowercase ) lowercase__ = generate_kwargs lowercase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.TENSORS if return_type is not None: lowercase__ = return_type if clean_up_tokenization_spaces is not None: lowercase__ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase ( self :int , *_lowercase :Optional[int] , **_lowercase :List[str] ): '''simple docstring''' if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self :Union[str, Any] , _lowercase :Dict , **_lowercase :Tuple ): '''simple docstring''' return super().__call__(_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple , _lowercase :Optional[int]="" , _lowercase :Tuple=None , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prompt_text if handle_long_generation == "hole": lowercase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: lowercase__ = generate_kwargs["max_new_tokens"] else: lowercase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: lowercase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) lowercase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: lowercase__ = inputs["attention_mask"][:, -keep_length:] return inputs def UpperCAmelCase ( self :str , _lowercase :int , **_lowercase :str ): '''simple docstring''' lowercase__ = model_inputs["input_ids"] lowercase__ = model_inputs.get("attention_mask" , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: lowercase__ = None lowercase__ = None lowercase__ = 1 else: lowercase__ = input_ids.shape[0] lowercase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowercase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: lowercase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: lowercase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowercase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowercase__ = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) lowercase__ = generated_sequence.shape[0] if self.framework == "pt": lowercase__ = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": lowercase__ = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCAmelCase ( self :Any , _lowercase :Tuple , _lowercase :str=ReturnType.FULL_TEXT , _lowercase :Dict=True ): '''simple docstring''' lowercase__ = model_outputs["generated_sequence"][0] lowercase__ = model_outputs["input_ids"] lowercase__ = model_outputs["prompt_text"] lowercase__ = generated_sequence.numpy().tolist() lowercase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowercase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowercase__ = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowercase__ = 0 else: lowercase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: lowercase__ = prompt_text + text[prompt_length:] else: lowercase__ = text[prompt_length:] lowercase__ = {"generated_text": all_text} records.append(_lowercase ) return records
655
1
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _snake_case = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _snake_case = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def _A ( __magic_name__ ): lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__magic_name__ )[0] @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) lowercase__ = data.reshape(__magic_name__ , __magic_name__ , __magic_name__ , 1 ) return data @deprecated(__magic_name__ , "Please use tf.one_hot on tensors." ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__magic_name__ ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=10 ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(__magic_name__ ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__magic_name__ , __magic_name__ ) return labels class lowerCAmelCase : @deprecated( _lowercase , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Tuple=False , _lowercase :str=False , _lowercase :Dict=dtypes.floataa , _lowercase :Optional[Any]=True , _lowercase :Any=None , ): '''simple docstring''' lowercase__ , lowercase__ = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowercase__ = 1_00_00 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'''images.shape: {images.shape} labels.shape: {labels.shape}''' lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(_lowercase , 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._images @property def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self._labels @property def UpperCAmelCase ( self :Dict ): '''simple docstring''' return self._num_examples @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._epochs_completed def UpperCAmelCase ( self :str , _lowercase :Union[str, Any] , _lowercase :Any=False , _lowercase :Union[str, Any]=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 7_84 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__magic_name__ , "Please write your own downloading logic." ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if not gfile.Exists(__magic_name__ ): gfile.MakeDirs(__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) if not gfile.Exists(__magic_name__ ): urllib.request.urlretrieve(__magic_name__ , __magic_name__ ) # noqa: S310 with gfile.GFile(__magic_name__ ) as f: lowercase__ = f.size() print("Successfully downloaded" , __magic_name__ , __magic_name__ , "bytes." ) return filepath @deprecated( __magic_name__ , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=False , __magic_name__=dtypes.floataa , __magic_name__=True , __magic_name__=5000 , __magic_name__=None , __magic_name__=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__magic_name__ , one_hot=__magic_name__ , dtype=__magic_name__ , seed=__magic_name__ ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = "train-images-idx3-ubyte.gz" lowercase__ = "train-labels-idx1-ubyte.gz" lowercase__ = "t10k-images-idx3-ubyte.gz" lowercase__ = "t10k-labels-idx1-ubyte.gz" lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) if not 0 <= validation_size <= len(__magic_name__ ): lowercase__ = ( "Validation size should be between 0 and " f'''{len(__magic_name__ )}. Received: {validation_size}.''' ) raise ValueError(__magic_name__ ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {"dtype": dtype, "reshape": reshape, "seed": seed} lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ )
655
1
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker _snake_case = """CompVis/stable-diffusion-v1-1""" _snake_case = """CompVis/stable-diffusion-v1-2""" _snake_case = """CompVis/stable-diffusion-v1-3""" _snake_case = """CompVis/stable-diffusion-v1-4""" class lowerCAmelCase ( lowercase_ ): def __init__( self :Tuple , _lowercase :AutoencoderKL , _lowercase :CLIPTextModel , _lowercase :CLIPTokenizer , _lowercase :UNetaDConditionModel , _lowercase :Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , _lowercase :StableDiffusionSafetyChecker , _lowercase :CLIPImageProcessor , _lowercase :bool = True , ): '''simple docstring''' super()._init_() lowercase__ = StableDiffusionPipeline.from_pretrained(_lowercase ) lowercase__ = StableDiffusionPipeline.from_pretrained(_lowercase ) lowercase__ = StableDiffusionPipeline.from_pretrained(_lowercase ) lowercase__ = StableDiffusionPipeline( vae=_lowercase , text_encoder=_lowercase , tokenizer=_lowercase , unet=_lowercase , scheduler=_lowercase , safety_checker=_lowercase , feature_extractor=_lowercase , requires_safety_checker=_lowercase , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def UpperCAmelCase ( self :int ): '''simple docstring''' return {k: getattr(self , _lowercase ) for k in self.config.keys() if not k.startswith("_" )} def UpperCAmelCase ( self :Dict , _lowercase :Optional[Union[str, int]] = "auto" ): '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowercase__ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_lowercase ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' self.enable_attention_slicing(_lowercase ) @torch.no_grad() def UpperCAmelCase ( self :Tuple , _lowercase :Union[str, List[str]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 50 , _lowercase :float = 7.5 , _lowercase :Optional[Union[str, List[str]]] = None , _lowercase :Optional[int] = 1 , _lowercase :float = 0.0 , _lowercase :Optional[torch.Generator] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , _lowercase :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowercase :int = 1 , **_lowercase :Dict , ): '''simple docstring''' return self.pipea( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) @torch.no_grad() def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 50 , _lowercase :float = 7.5 , _lowercase :Optional[Union[str, List[str]]] = None , _lowercase :Optional[int] = 1 , _lowercase :float = 0.0 , _lowercase :Optional[torch.Generator] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , _lowercase :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowercase :int = 1 , **_lowercase :List[str] , ): '''simple docstring''' return self.pipea( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) @torch.no_grad() def UpperCAmelCase ( self :Any , _lowercase :Union[str, List[str]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 50 , _lowercase :float = 7.5 , _lowercase :Optional[Union[str, List[str]]] = None , _lowercase :Optional[int] = 1 , _lowercase :float = 0.0 , _lowercase :Optional[torch.Generator] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , _lowercase :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowercase :int = 1 , **_lowercase :Optional[Any] , ): '''simple docstring''' return self.pipea( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) @torch.no_grad() def UpperCAmelCase ( self :List[Any] , _lowercase :Union[str, List[str]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 50 , _lowercase :float = 7.5 , _lowercase :Optional[Union[str, List[str]]] = None , _lowercase :Optional[int] = 1 , _lowercase :float = 0.0 , _lowercase :Optional[torch.Generator] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , _lowercase :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowercase :int = 1 , **_lowercase :Optional[int] , ): '''simple docstring''' return self.pipea( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) @torch.no_grad() def UpperCAmelCase ( self :List[str] , _lowercase :Union[str, List[str]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 50 , _lowercase :float = 7.5 , _lowercase :Optional[Union[str, List[str]]] = None , _lowercase :Optional[int] = 1 , _lowercase :float = 0.0 , _lowercase :Optional[torch.Generator] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , _lowercase :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowercase :int = 1 , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = "cuda" if torch.cuda.is_available() else "cpu" self.to(_lowercase ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` must be divisible by 8 but are {height} and {width}.''' ) # Get first result from Stable Diffusion Checkpoint v1.1 lowercase__ = self.textaimg_sda_a( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.2 lowercase__ = self.textaimg_sda_a( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.3 lowercase__ = self.textaimg_sda_a( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) # Get first result from Stable Diffusion Checkpoint v1.4 lowercase__ = self.textaimg_sda_a( prompt=_lowercase , height=_lowercase , width=_lowercase , num_inference_steps=_lowercase , guidance_scale=_lowercase , negative_prompt=_lowercase , num_images_per_prompt=_lowercase , eta=_lowercase , generator=_lowercase , latents=_lowercase , output_type=_lowercase , return_dict=_lowercase , callback=_lowercase , callback_steps=_lowercase , **_lowercase , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
655
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { """configuration_layoutlmv3""": [ """LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LayoutLMv3Config""", """LayoutLMv3OnnxConfig""", ], """processing_layoutlmv3""": ["""LayoutLMv3Processor"""], """tokenization_layoutlmv3""": ["""LayoutLMv3Tokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ["""LayoutLMv3TokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST""", """LayoutLMv3ForQuestionAnswering""", """LayoutLMv3ForSequenceClassification""", """LayoutLMv3ForTokenClassification""", """LayoutLMv3Model""", """LayoutLMv3PreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFLayoutLMv3ForQuestionAnswering""", """TFLayoutLMv3ForSequenceClassification""", """TFLayoutLMv3ForTokenClassification""", """TFLayoutLMv3Model""", """TFLayoutLMv3PreTrainedModel""", ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ["""LayoutLMv3FeatureExtractor"""] _snake_case = ["""LayoutLMv3ImageProcessor"""] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
1
def _A ( __magic_name__ ): lowercase__ = set() # edges = list of graph's edges lowercase__ = get_edges(__magic_name__ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowercase__ , lowercase__ = edges.pop() chosen_vertices.add(__magic_name__ ) chosen_vertices.add(__magic_name__ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(__magic_name__ ) return chosen_vertices def _A ( __magic_name__ ): lowercase__ = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
655
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["prompt"] lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] if "image" in inputs: lowercase__ = inputs["image"] else: lowercase__ = None if "mask_image" in inputs: lowercase__ = inputs["mask_image"] else: lowercase__ = None if "original_image" in inputs: lowercase__ = inputs["original_image"] else: lowercase__ = None lowercase__ , lowercase__ = pipe.encode_prompt(_lowercase ) # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_lowercase , _lowercase , _lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowercase , _lowercase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 )
655
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter _snake_case = """Create a default config file for Accelerate with only a few flags set.""" def _A ( __magic_name__="no" , __magic_name__ = default_json_config_file , __magic_name__ = False ): lowercase__ = Path(__magic_name__ ) path.parent.mkdir(parents=__magic_name__ , exist_ok=__magic_name__ ) if path.exists(): print( f'''Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.''' ) return False lowercase__ = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( f'''`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}''' ) lowercase__ = { "compute_environment": "LOCAL_MACHINE", "mixed_precision": mixed_precision, } if torch.cuda.is_available(): lowercase__ = torch.cuda.device_count() lowercase__ = num_gpus lowercase__ = False if num_gpus > 1: lowercase__ = "MULTI_GPU" else: lowercase__ = "NO" elif is_xpu_available() and use_xpu: lowercase__ = torch.xpu.device_count() lowercase__ = num_xpus lowercase__ = False if num_xpus > 1: lowercase__ = "MULTI_XPU" else: lowercase__ = "NO" elif is_npu_available(): lowercase__ = torch.npu.device_count() lowercase__ = num_npus lowercase__ = False if num_npus > 1: lowercase__ = "MULTI_NPU" else: lowercase__ = "NO" else: lowercase__ = 0 lowercase__ = True lowercase__ = 1 lowercase__ = "NO" lowercase__ = ClusterConfig(**__magic_name__ ) config.to_json_file(__magic_name__ ) return path def _A ( __magic_name__ , __magic_name__ ): lowercase__ = parser.add_parser("default" , parents=__magic_name__ , help=__magic_name__ , formatter_class=__magic_name__ ) parser.add_argument( "--config_file" , default=__magic_name__ , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , dest="save_location" , ) parser.add_argument( "--mixed_precision" , choices=["no", "fp16", "bf16"] , type=__magic_name__ , help="Whether or not to use mixed precision training. " "Choose between FP16 and BF16 (bfloat16) training. " "BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later." , default="no" , ) parser.set_defaults(func=__magic_name__ ) return parser def _A ( __magic_name__ ): lowercase__ = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(f'''accelerate configuration saved at {config_file}''' )
655
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) lowercase__ = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase__ = model(_lowercase )["last_hidden_state"] lowercase__ = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
655
1
import baseaa def _A ( __magic_name__ ): return baseaa.baaencode(string.encode("utf-8" ) ) def _A ( __magic_name__ ): return baseaa.baadecode(__magic_name__ ).decode("utf-8" ) if __name__ == "__main__": _snake_case = """Hello World!""" _snake_case = baseaa_encode(test) print(encoded) _snake_case = baseaa_decode(encoded) print(decoded)
655
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
1
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _snake_case = logging.get_logger(__name__) _snake_case = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} # See all LED models at https://huggingface.co/models?filter=LED _snake_case = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } _snake_case = { """allenai/led-base-16384""": 1_6384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def _A ( ): lowercase__ = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) lowercase__ = bs[:] lowercase__ = 0 for b in range(2**8 ): if b not in bs: bs.append(__magic_name__ ) cs.append(2**8 + n ) n += 1 lowercase__ = [chr(__magic_name__ ) for n in cs] return dict(zip(__magic_name__ , __magic_name__ ) ) def _A ( __magic_name__ ): lowercase__ = set() lowercase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ = char return pairs class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = ['input_ids', 'attention_mask'] def __init__( self :Tuple , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :Optional[Any]="replace" , _lowercase :Dict="<s>" , _lowercase :Tuple="</s>" , _lowercase :int="</s>" , _lowercase :List[str]="<s>" , _lowercase :Optional[int]="<unk>" , _lowercase :Tuple="<pad>" , _lowercase :Union[str, Any]="<mask>" , _lowercase :int=False , **_lowercase :int , ): '''simple docstring''' lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else bos_token lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else eos_token lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else sep_token lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else cls_token lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else unk_token lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else mask_token super().__init__( errors=_lowercase , bos_token=_lowercase , eos_token=_lowercase , unk_token=_lowercase , sep_token=_lowercase , cls_token=_lowercase , pad_token=_lowercase , mask_token=_lowercase , add_prefix_space=_lowercase , **_lowercase , ) with open(_lowercase , encoding="utf-8" ) as vocab_handle: lowercase__ = json.load(_lowercase ) lowercase__ = {v: k for k, v in self.encoder.items()} lowercase__ = errors # how to handle errors in decoding lowercase__ = bytes_to_unicode() lowercase__ = {v: k for k, v in self.byte_encoder.items()} with open(_lowercase , encoding="utf-8" ) as merges_handle: lowercase__ = merges_handle.read().split("\n" )[1:-1] lowercase__ = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) lowercase__ = {} lowercase__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase ( self :int ): '''simple docstring''' return len(self.encoder ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def UpperCAmelCase ( self :Any , _lowercase :Optional[Any] ): '''simple docstring''' if token in self.cache: return self.cache[token] lowercase__ = tuple(_lowercase ) lowercase__ = get_pairs(_lowercase ) if not pairs: return token while True: lowercase__ = min(_lowercase , key=lambda _lowercase : self.bpe_ranks.get(_lowercase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ = bigram lowercase__ = [] lowercase__ = 0 while i < len(_lowercase ): try: lowercase__ = word.index(_lowercase , _lowercase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ = j if word[i] == first and i < len(_lowercase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ = tuple(_lowercase ) lowercase__ = new_word if len(_lowercase ) == 1: break else: lowercase__ = get_pairs(_lowercase ) lowercase__ = " ".join(_lowercase ) lowercase__ = word return word def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = [] for token in re.findall(self.pat , _lowercase ): lowercase__ = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_lowercase ).split(" " ) ) return bpe_tokens def UpperCAmelCase ( self :Union[str, Any] , _lowercase :str ): '''simple docstring''' return self.encoder.get(_lowercase , self.encoder.get(self.unk_token ) ) def UpperCAmelCase ( self :Dict , _lowercase :Tuple ): '''simple docstring''' return self.decoder.get(_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = "".join(_lowercase ) lowercase__ = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def UpperCAmelCase ( self :int , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_lowercase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_lowercase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowercase , ensure_ascii=_lowercase ) + "\n" ) lowercase__ = 0 with open(_lowercase , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowercase__ = token_index writer.write(" ".join(_lowercase ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase ( self :Optional[Any] , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase ( self :Dict , _lowercase :List[int] , _lowercase :Optional[List[int]] = None , _lowercase :bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowercase , token_ids_a=_lowercase , already_has_special_tokens=_lowercase ) if token_ids_a is None: return [1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase ( self :Union[str, Any] , _lowercase :List[str] , _lowercase :Union[str, Any]=False , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_lowercase ) > 0 and not text[0].isspace()): lowercase__ = " " + text return (text, kwargs) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Union[Dict[str, EncodedInput], BatchEncoding] , _lowercase :Optional[int] = None , _lowercase :PaddingStrategy = PaddingStrategy.DO_NOT_PAD , _lowercase :Optional[int] = None , _lowercase :Optional[bool] = None , ): '''simple docstring''' lowercase__ = super()._pad( encoded_inputs=_lowercase , max_length=_lowercase , padding_strategy=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , ) # Load from model defaults if return_attention_mask is None: lowercase__ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ = len(encoded_inputs["global_attention_mask"] ) != len(_lowercase ) if needs_to_be_padded: lowercase__ = len(_lowercase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
655
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
1
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def _A ( __magic_name__ ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__magic_name__ , "_dynamo" ): return False return isinstance(__magic_name__ , torch._dynamo.eval_frame.OptimizedModule ) def _A ( __magic_name__ , __magic_name__ = True ): lowercase__ = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) lowercase__ = is_compiled_module(__magic_name__ ) if is_compiled: lowercase__ = model lowercase__ = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__magic_name__ , __magic_name__ ): lowercase__ = model.module if not keep_fpaa_wrapper: lowercase__ = getattr(__magic_name__ , "forward" ) lowercase__ = model.__dict__.pop("_original_forward" , __magic_name__ ) if original_forward is not None: while hasattr(__magic_name__ , "__wrapped__" ): lowercase__ = forward.__wrapped__ if forward == original_forward: break lowercase__ = forward if getattr(__magic_name__ , "_converted_to_transformer_engine" , __magic_name__ ): convert_model(__magic_name__ , to_transformer_engine=__magic_name__ ) if is_compiled: lowercase__ = model lowercase__ = compiled_model return model def _A ( ): PartialState().wait_for_everyone() def _A ( __magic_name__ , __magic_name__ ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__magic_name__ , __magic_name__ ) elif PartialState().local_process_index == 0: torch.save(__magic_name__ , __magic_name__ ) @contextmanager def _A ( **__magic_name__ ): for key, value in kwargs.items(): lowercase__ = str(__magic_name__ ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def _A ( __magic_name__ ): if not hasattr(__magic_name__ , "__qualname__" ) and not hasattr(__magic_name__ , "__name__" ): lowercase__ = getattr(__magic_name__ , "__class__" , __magic_name__ ) if hasattr(__magic_name__ , "__qualname__" ): return obj.__qualname__ if hasattr(__magic_name__ , "__name__" ): return obj.__name__ return str(__magic_name__ ) def _A ( __magic_name__ , __magic_name__ ): for key, value in source.items(): if isinstance(__magic_name__ , __magic_name__ ): lowercase__ = destination.setdefault(__magic_name__ , {} ) merge_dicts(__magic_name__ , __magic_name__ ) else: lowercase__ = value return destination def _A ( __magic_name__ = None ): if port is None: lowercase__ = 2_9500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
655
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = SMALL_MODEL_IDENTIFIER lowercase__ = "pt" lowercase__ = "tf" def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :int ): '''simple docstring''' lowercase__ = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = "mock_framework" # Framework provided - return whatever the user provides lowercase__ = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model )
655
1
import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) _snake_case = logging.getLogger() def _A ( __magic_name__ , __magic_name__ ): lowercase__ = "\n".join(__magic_name__ ) Path(__magic_name__ ).open("w" ).writelines(__magic_name__ ) _snake_case = """patrickvonplaten/t5-tiny-random""" _snake_case = """sshleifer/bart-tiny-random""" _snake_case = """sshleifer/tiny-mbart""" _snake_case = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Any , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" lowercase__ = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() lowercase__ = [" New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County."] _dump_articles(_lowercase , _lowercase ) lowercase__ = str(Path(self.get_auto_remove_tmp_dir() ) / "scores.json" ) lowercase__ = "translation_en_to_de" if model == T5_TINY else "summarization" lowercase__ = f''' run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 '''.split() with patch.object(_lowercase , "argv" , _lowercase ): run_generate() assert Path(_lowercase ).exists() # os.remove(Path(output_file_name)) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' self.run_eval_tester(_lowercase ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[Any] ): '''simple docstring''' self.run_eval_tester(_lowercase ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def UpperCAmelCase ( self :Union[str, Any] , _lowercase :int ): '''simple docstring''' lowercase__ = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" lowercase__ = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() lowercase__ = { "en": ["Machine learning is great, isn't it?", "I like to eat bananas", "Tomorrow is another great day!"], "de": [ "Maschinelles Lernen ist großartig, oder?", "Ich esse gerne Bananen", "Morgen ist wieder ein toller Tag!", ], } lowercase__ = Path(self.get_auto_remove_tmp_dir() ) lowercase__ = str(tmp_dir / "scores.json" ) lowercase__ = str(tmp_dir / "val.target" ) _dump_articles(_lowercase , text["en"] ) _dump_articles(_lowercase , text["de"] ) lowercase__ = "translation_en_to_de" if model == T5_TINY else "summarization" lowercase__ = f''' run_eval_search.py {model} {str(_lowercase )} {str(_lowercase )} --score_path {score_path} --reference_path {reference_path} --task {task} '''.split() testargs.extend(["--search", "num_beams=1:2 length_penalty=0.9:1.0"] ) with patch.object(_lowercase , "argv" , _lowercase ): with CaptureStdout() as cs: run_search() lowercase__ = [" num_beams | length_penalty", model, "Best score args"] lowercase__ = ["Info"] if "translation" in task: expected_strings.append("bleu" ) else: expected_strings.extend(_lowercase ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(_lowercase ).exists() os.remove(Path(_lowercase ) )
655
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git_vision_model' def __init__( self :Dict , _lowercase :Dict=7_68 , _lowercase :Dict=30_72 , _lowercase :Tuple=12 , _lowercase :List[str]=12 , _lowercase :Tuple=3 , _lowercase :Dict=2_24 , _lowercase :Tuple=16 , _lowercase :Optional[int]="quick_gelu" , _lowercase :Union[str, Any]=1e-5 , _lowercase :Tuple=0.0 , _lowercase :Tuple=0.02 , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = hidden_size lowercase__ = intermediate_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_channels lowercase__ = patch_size lowercase__ = image_size lowercase__ = initializer_range lowercase__ = attention_dropout lowercase__ = layer_norm_eps lowercase__ = hidden_act @classmethod def UpperCAmelCase ( cls :List[str] , _lowercase :Union[str, os.PathLike] , **_lowercase :Optional[int] ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git' def __init__( self :Union[str, Any] , _lowercase :Dict=None , _lowercase :List[str]=3_05_22 , _lowercase :Tuple=7_68 , _lowercase :Any=6 , _lowercase :Dict=12 , _lowercase :Any=30_72 , _lowercase :List[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :Optional[int]=0.1 , _lowercase :Optional[Any]=10_24 , _lowercase :Any=0.02 , _lowercase :int=1e-12 , _lowercase :List[Any]=0 , _lowercase :int="absolute" , _lowercase :List[str]=True , _lowercase :Any=False , _lowercase :int=1_01 , _lowercase :str=1_02 , _lowercase :Dict=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , pad_token_id=_lowercase , **_lowercase ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowercase__ = GitVisionConfig(**_lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = tie_word_embeddings lowercase__ = num_image_with_embedding lowercase__ = bos_token_id lowercase__ = eos_token_id def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
655
1
from pathlib import Path import fire from tqdm import tqdm def _A ( __magic_name__="ro" , __magic_name__="en" , __magic_name__="wmt16" , __magic_name__=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("run pip install datasets" ) lowercase__ = f'''{src_lang}-{tgt_lang}''' print(f'''Converting {dataset}-{pair}''' ) lowercase__ = datasets.load_dataset(__magic_name__ , __magic_name__ ) if save_dir is None: lowercase__ = f'''{dataset}-{pair}''' lowercase__ = Path(__magic_name__ ) save_dir.mkdir(exist_ok=__magic_name__ ) for split in ds.keys(): print(f'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets lowercase__ = "val" if split == "validation" else split lowercase__ = save_dir.joinpath(f'''{fn}.source''' ) lowercase__ = save_dir.joinpath(f'''{fn}.target''' ) lowercase__ = src_path.open("w+" ) lowercase__ = tgt_path.open("w+" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): lowercase__ = x["translation"] src_fp.write(ex[src_lang] + "\n" ) tgt_fp.write(ex[tgt_lang] + "\n" ) print(f'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
655
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
1
import os import jsonlines import numpy as np from tqdm import tqdm _snake_case = 2048 _snake_case = 4096 _snake_case = 42 _snake_case = os.environ.pop("""PROCESS_TRAIN""", """false""") _snake_case = {"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def _A ( __magic_name__ ): def choose_first(__magic_name__ , __magic_name__=False ): assert isinstance(__magic_name__ , __magic_name__ ) if len(__magic_name__ ) == 1: lowercase__ = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: lowercase__ = {k: [a[k]] for k in a} if len(a["start_token"] ) > 0: break return a lowercase__ = {"id": example["id"]} lowercase__ = example["annotations"] lowercase__ = annotation["yes_no_answer"] if 0 in yes_no_answer or 1 in yes_no_answer: lowercase__ = ["yes"] if 1 in yes_no_answer else ["no"] lowercase__ = lowercase__ = [] lowercase__ = lowercase__ = [] lowercase__ = ["<cls>"] else: lowercase__ = ["short"] lowercase__ = choose_first(annotation["short_answers"] ) if len(out["start_token"] ) == 0: # answer will be long if short is not available lowercase__ = ["long"] lowercase__ = choose_first(annotation["long_answer"] , is_long_answer=__magic_name__ ) lowercase__ = [] answer.update(__magic_name__ ) # disregard some samples if len(answer["start_token"] ) > 1 or answer["start_token"] == answer["end_token"]: lowercase__ = True else: lowercase__ = False lowercase__ = ["start_token", "end_token", "start_byte", "end_byte", "text"] if not all(isinstance(answer[k] , __magic_name__ ) for k in cols ): raise ValueError("Issue in ID" , example["id"] ) return answer def _A ( __magic_name__ , __magic_name__=False ): lowercase__ = _get_single_answer(__magic_name__ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element lowercase__ = example["document"]["tokens"] lowercase__ = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) return { "context": " ".join(__magic_name__ ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples lowercase__ = ["start_token", "end_token"] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 lowercase__ = example["document"]["tokens"] lowercase__ = answer["start_token"] lowercase__ = answer["end_token"] lowercase__ = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 lowercase__ = " ".join(context[start_token:end_token] ) # checking above code if assertion: lowercase__ = doc["is_html"][answer["start_token"] : answer["end_token"]] lowercase__ = doc["token"][answer["start_token"] : answer["end_token"]] lowercase__ = " ".join([old[i] for i in range(len(__magic_name__ ) ) if not is_html[i]] ) if new != old: print("ID:" , example["id"] ) print("New:" , __magic_name__ , end="\n" ) print("Old:" , __magic_name__ , end="\n\n" ) return { "context": " ".join(__magic_name__ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def _A ( __magic_name__ , __magic_name__ , __magic_name__=2048 , __magic_name__=4096 , __magic_name__=True ): # overlap will be of doc_stride - q_len lowercase__ = get_context_and_ans(__magic_name__ , assertion=__magic_name__ ) lowercase__ = out["answer"] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } lowercase__ = tokenizer(example["question"]["text"] , out["context"] ).input_ids lowercase__ = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element lowercase__ = [] lowercase__ = [] lowercase__ = input_ids[:q_len] lowercase__ = range(__magic_name__ , len(__magic_name__ ) , max_length - doc_stride ) for i in doc_start_indices: lowercase__ = i + max_length - q_len lowercase__ = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer["category"][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(__magic_name__ ), "end_token": [-100] * len(__magic_name__ ), "category": category, }, } lowercase__ = out["context"].split() lowercase__ = splitted_context[answer["end_token"]] lowercase__ = len( tokenizer( " ".join(splitted_context[: answer["start_token"]] ) , add_special_tokens=__magic_name__ , ).input_ids ) lowercase__ = len( tokenizer(" ".join(splitted_context[: answer["end_token"]] ) , add_special_tokens=__magic_name__ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token lowercase__ = len(tokenizer(__magic_name__ , add_special_tokens=__magic_name__ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 lowercase__ = input_ids[answer["start_token"] : answer["end_token"] + 1] # right & left are inclusive lowercase__ = answer["start_token"] lowercase__ = answer["end_token"] if assertion: lowercase__ = tokenizer.decode(__magic_name__ ) if answer["span"] != new: print("ISSUE IN TOKENIZATION" ) print("OLD:" , answer["span"] ) print("NEW:" , __magic_name__ , end="\n\n" ) if len(__magic_name__ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } lowercase__ = input_ids[:q_len] lowercase__ = range(__magic_name__ , len(__magic_name__ ) , max_length - doc_stride ) lowercase__ = [] lowercase__ = [] lowercase__ = [] lowercase__ = [] # null, yes, no, long, short for i in doc_start_indices: lowercase__ = i + max_length - q_len lowercase__ = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: lowercase__ = start_token - i + q_len lowercase__ = end_token - i + q_len answers_category.append(answer["category"][0] ) # ["short"] -> "short" else: lowercase__ = -100 lowercase__ = -100 answers_category.append("null" ) lowercase__ = inputs[-1][start_token : end_token + 1] answers_start_token.append(__magic_name__ ) answers_end_token.append(__magic_name__ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print("ISSUE in strided for ID:" , example["id"] ) print("New:" , tokenizer.decode(__magic_name__ ) ) print("Old:" , tokenizer.decode(__magic_name__ ) , end="\n\n" ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def _A ( __magic_name__ , __magic_name__ , __magic_name__=2048 , __magic_name__=4096 , __magic_name__=False ): lowercase__ = get_strided_contexts_and_ans( __magic_name__ , __magic_name__ , doc_stride=__magic_name__ , max_length=__magic_name__ , assertion=__magic_name__ , ) return example def _A ( __magic_name__ , __magic_name__ ): with jsonlines.open(__magic_name__ , "a" ) as writer: for example in tqdm(__magic_name__ , total=len(__magic_name__ ) , desc="Saving samples ... " ): lowercase__ = example["labels"] for ids, start, end, cat in zip( example["input_ids"] , labels["start_token"] , labels["end_token"] , labels["category"] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { "input_ids": ids, "start_token": start, "end_token": end, "category": CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer _snake_case = load_dataset("""natural_questions""") _snake_case = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") _snake_case = data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] _snake_case = { """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } _snake_case = data.map(prepare_inputs, fn_kwargs=fn_kwargs) _snake_case = data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) _snake_case = """nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
655
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _snake_case = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val def _A ( __magic_name__ ): lowercase__ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowercase__ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) lowercase__ = value else: lowercase__ = value return new_state_dict def _A ( __magic_name__ ): lowercase__ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowercase__ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowercase__ = in_proj_weight_cross_attn[:256, :] lowercase__ = in_proj_bias_cross_attn[:256] lowercase__ = in_proj_weight_cross_attn[256:512, :] lowercase__ = in_proj_bias_cross_attn[256:512] lowercase__ = in_proj_weight_cross_attn[-256:, :] lowercase__ = in_proj_bias_cross_attn[-256:] def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = image.size lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 800 if "detection" in checkpoint_url else 1000 lowercase__ = target_max_size / current_max_size lowercase__ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _A ( __magic_name__ ): lowercase__ = F.to_tensor(__magic_name__ ) lowercase__ = F.normalize(__magic_name__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): logger.info("Converting model..." ) # load original state dict lowercase__ = torch.hub.load_state_dict_from_url(__magic_name__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowercase__ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val # create HuggingFace model and load state dict lowercase__ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: lowercase__ = 15 lowercase__ = 2 lowercase__ = {0: "table", 1: "table rotated"} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 125 lowercase__ = 6 lowercase__ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) lowercase__ = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion lowercase__ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" lowercase__ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=__magic_name__ ) lowercase__ = Image.open(__magic_name__ ).convert("RGB" ) lowercase__ = normalize(resize(__magic_name__ , __magic_name__ ) ).unsqueeze(0 ) lowercase__ = model(__magic_name__ ) if "detection" in checkpoint_url: lowercase__ = (1, 15, 3) lowercase__ = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) lowercase__ = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: lowercase__ = (1, 125, 7) lowercase__ = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) lowercase__ = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __magic_name__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) lowercase__ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _snake_case = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
655
1
from collections.abc import Generator from math import sin def _A ( __magic_name__ ): if len(__magic_name__ ) != 32: raise ValueError("Input must be of length 32" ) lowercase__ = B"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def _A ( __magic_name__ ): if i < 0: raise ValueError("Input must be non-negative" ) lowercase__ = format(__magic_name__ , "08x" )[-8:] lowercase__ = B"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def _A ( __magic_name__ ): lowercase__ = B"" for char in message: bit_string += format(__magic_name__ , "08b" ).encode("utf-8" ) lowercase__ = format(len(__magic_name__ ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__magic_name__ ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def _A ( __magic_name__ ): if len(__magic_name__ ) % 512 != 0: raise ValueError("Input must have length that's a multiple of 512" ) for pos in range(0 , len(__magic_name__ ) , 512 ): lowercase__ = bit_string[pos : pos + 512] lowercase__ = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def _A ( __magic_name__ ): if i < 0: raise ValueError("Input must be non-negative" ) lowercase__ = format(__magic_name__ , "032b" ) lowercase__ = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(__magic_name__ , 2 ) def _A ( __magic_name__ , __magic_name__ ): return (a + b) % 2**32 def _A ( __magic_name__ , __magic_name__ ): if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def _A ( __magic_name__ ): lowercase__ = preprocess(__magic_name__ ) lowercase__ = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states lowercase__ = 0X6745_2301 lowercase__ = 0Xefcd_ab89 lowercase__ = 0X98ba_dcfe lowercase__ = 0X1032_5476 lowercase__ = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__magic_name__ ): lowercase__ = aa lowercase__ = ba lowercase__ = ca lowercase__ = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowercase__ = d ^ (b & (c ^ d)) lowercase__ = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowercase__ = c ^ (d & (b ^ c)) lowercase__ = (5 * i + 1) % 16 elif i <= 47: lowercase__ = b ^ c ^ d lowercase__ = (3 * i + 5) % 16 else: lowercase__ = c ^ (b | not_aa(__magic_name__ )) lowercase__ = (7 * i) % 16 lowercase__ = (f + a + added_consts[i] + block_words[g]) % 2**32 lowercase__ = d lowercase__ = c lowercase__ = b lowercase__ = sum_aa(__magic_name__ , left_rotate_aa(__magic_name__ , shift_amounts[i] ) ) # Add hashed chunk to running total lowercase__ = sum_aa(__magic_name__ , __magic_name__ ) lowercase__ = sum_aa(__magic_name__ , __magic_name__ ) lowercase__ = sum_aa(__magic_name__ , __magic_name__ ) lowercase__ = sum_aa(__magic_name__ , __magic_name__ ) lowercase__ = reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) return digest if __name__ == "__main__": import doctest doctest.testmod()
655
from typing import TYPE_CHECKING from ...utils import _LazyModule _snake_case = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
1
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = IFImgaImgSuperResolutionPipeline __lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} __lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'original_image'} ) __lowerCamelCase = PipelineTesterMixin.required_optional_params - {'latents'} def UpperCAmelCase ( self :int ): '''simple docstring''' return self._get_superresolution_dummy_components() def UpperCAmelCase ( self :int , _lowercase :Dict , _lowercase :Optional[Any]=0 ): '''simple docstring''' if str(_lowercase ).startswith("mps" ): lowercase__ = torch.manual_seed(_lowercase ) else: lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowercase ) ).to(_lowercase ) lowercase__ = floats_tensor((1, 3, 16, 16) , rng=random.Random(_lowercase ) ).to(_lowercase ) lowercase__ = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCAmelCase ( self :Any ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' self._test_save_load_local() def UpperCAmelCase ( self :str ): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
655
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = """ Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _A ( __magic_name__ , __magic_name__ , __magic_name__=8 ): lowercase__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :UNetaDConditionModel , _lowercase :DDPMScheduler , _lowercase :VQModel , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) lowercase__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :int , _lowercase :str ): '''simple docstring''' if latents is None: lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase__ = latents.to(_lowercase ) lowercase__ = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self :int , _lowercase :int=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) lowercase__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase__ = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase__ , lowercase__ = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. lowercase__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__( self :int , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 1_00 , _lowercase :float = 4.0 , _lowercase :int = 1 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = self._execution_device lowercase__ = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) lowercase__ = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: lowercase__ = image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) lowercase__ = self.scheduler.timesteps lowercase__ = self.unet.config.in_channels lowercase__ , lowercase__ = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) # create initial latent lowercase__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowercase , _lowercase , _lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance lowercase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase__ = {"image_embeds": image_embeds} lowercase__ = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) lowercase__ , lowercase__ = noise_pred.chunk(2 ) lowercase__ , lowercase__ = variance_pred.chunk(2 ) lowercase__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase__ = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing lowercase__ = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase__ = image * 0.5 + 0.5 lowercase__ = image.clamp(0 , 1 ) lowercase__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
655
1
from __future__ import annotations import bisect def _A ( __magic_name__ , __magic_name__ , __magic_name__ = 0 , __magic_name__ = -1 ): if hi < 0: lowercase__ = len(__magic_name__ ) while lo < hi: lowercase__ = lo + (hi - lo) // 2 if sorted_collection[mid] < item: lowercase__ = mid + 1 else: lowercase__ = mid return lo def _A ( __magic_name__ , __magic_name__ , __magic_name__ = 0 , __magic_name__ = -1 ): if hi < 0: lowercase__ = len(__magic_name__ ) while lo < hi: lowercase__ = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: lowercase__ = mid + 1 else: lowercase__ = mid return lo def _A ( __magic_name__ , __magic_name__ , __magic_name__ = 0 , __magic_name__ = -1 ): sorted_collection.insert(bisect_left(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) , __magic_name__ ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ = 0 , __magic_name__ = -1 ): sorted_collection.insert(bisect_right(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) , __magic_name__ ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = 0 lowercase__ = len(__magic_name__ ) - 1 while left <= right: lowercase__ = left + (right - left) // 2 lowercase__ = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: lowercase__ = midpoint - 1 else: lowercase__ = midpoint + 1 return None def _A ( __magic_name__ , __magic_name__ ): lowercase__ = bisect.bisect_left(__magic_name__ , __magic_name__ ) if index != len(__magic_name__ ) and sorted_collection[index] == item: return index return None def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): if right < left: return None lowercase__ = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(__magic_name__ , __magic_name__ , __magic_name__ , midpoint - 1 ) else: return binary_search_by_recursion(__magic_name__ , __magic_name__ , midpoint + 1 , __magic_name__ ) if __name__ == "__main__": _snake_case = input("""Enter numbers separated by comma:\n""").strip() _snake_case = sorted(int(item) for item in user_input.split(""",""")) _snake_case = int(input("""Enter a single number to be found in the list:\n""")) _snake_case = binary_search(collection, target) if result is None: print(F"""{target} was not found in {collection}.""") else: print(F"""{target} was found at position {result} in {collection}.""")
655
import inspect import unittest class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :int ): '''simple docstring''' try: import diffusers # noqa: F401 except ImportError: assert False def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' import diffusers from diffusers.dependency_versions_table import deps lowercase__ = inspect.getmembers(_lowercase , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowercase__ = "k-diffusion" elif backend == "invisible_watermark": lowercase__ = "invisible-watermark" assert backend in deps, f'''{backend} is not in the deps table!'''
655
1
from math import sqrt def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and ( number >= 0 ), "'number' must been an int and positive" lowercase__ = True # 0 and 1 are none primes. if number <= 1: lowercase__ = False for divisor in range(2 , int(round(sqrt(__magic_name__ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowercase__ = False break # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'status' must been from type bool" return status def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowercase__ = list(range(2 , n + 1 ) ) lowercase__ = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__magic_name__ ) ): for j in range(i + 1 , len(__magic_name__ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowercase__ = 0 # filters actual prime numbers. lowercase__ = [x for x in begin_list if x != 0] # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'ans' must been from type list" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n > 2), "'N' must been an int and > 2" lowercase__ = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(__magic_name__ ): ans.append(__magic_name__ ) # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'ans' must been from type list" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and number >= 0, "'number' must been an int and >= 0" lowercase__ = [] # this list will be returns of the function. # potential prime number factors. lowercase__ = 2 lowercase__ = number if number == 0 or number == 1: ans.append(__magic_name__ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__magic_name__ ): while quotient != 1: if is_prime(__magic_name__ ) and (quotient % factor == 0): ans.append(__magic_name__ ) quotient /= factor else: factor += 1 else: ans.append(__magic_name__ ) # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'ans' must been from type list" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ = 0 # prime factorization of 'number' lowercase__ = prime_factorization(__magic_name__ ) lowercase__ = max(__magic_name__ ) # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'ans' must been from type int" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ = 0 # prime factorization of 'number' lowercase__ = prime_factorization(__magic_name__ ) lowercase__ = min(__magic_name__ ) # precondition assert isinstance(__magic_name__ , __magic_name__ ), "'ans' must been from type int" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ), "'number' must been an int" assert isinstance(number % 2 == 0 , __magic_name__ ), "compare bust been from type bool" return number % 2 == 0 def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ), "'number' must been an int" assert isinstance(number % 2 != 0 , __magic_name__ ), "compare bust been from type bool" return number % 2 != 0 def _A ( __magic_name__ ): assert ( isinstance(__magic_name__ , __magic_name__ ) and (number > 2) and is_even(__magic_name__ ) ), "'number' must been an int, even and > 2" lowercase__ = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowercase__ = get_prime_numbers(__magic_name__ ) lowercase__ = len(__magic_name__ ) # run variable for while-loops. lowercase__ = 0 lowercase__ = None # exit variable. for break up the loops lowercase__ = True while i < len_pn and loop: lowercase__ = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowercase__ = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__magic_name__ , __magic_name__ ) and (len(__magic_name__ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def _A ( __magic_name__ , __magic_name__ ): assert ( isinstance(__magic_name__ , __magic_name__ ) and isinstance(__magic_name__ , __magic_name__ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowercase__ = 0 while numbera != 0: lowercase__ = numbera % numbera lowercase__ = numbera lowercase__ = rest # precondition assert isinstance(__magic_name__ , __magic_name__ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def _A ( __magic_name__ , __magic_name__ ): assert ( isinstance(__magic_name__ , __magic_name__ ) and isinstance(__magic_name__ , __magic_name__ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowercase__ = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowercase__ = prime_factorization(__magic_name__ ) lowercase__ = prime_factorization(__magic_name__ ) elif numbera == 1 or numbera == 1: lowercase__ = [] lowercase__ = [] lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 0 lowercase__ = 0 lowercase__ = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowercase__ = prime_fac_a.count(__magic_name__ ) lowercase__ = prime_fac_a.count(__magic_name__ ) for _ in range(max(__magic_name__ , __magic_name__ ) ): ans *= n else: lowercase__ = prime_fac_a.count(__magic_name__ ) for _ in range(__magic_name__ ): ans *= n done.append(__magic_name__ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowercase__ = prime_fac_a.count(__magic_name__ ) for _ in range(__magic_name__ ): ans *= n done.append(__magic_name__ ) # precondition assert isinstance(__magic_name__ , __magic_name__ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n >= 0), "'number' must been a positive int" lowercase__ = 0 lowercase__ = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__magic_name__ ): ans += 1 # precondition assert isinstance(__magic_name__ , __magic_name__ ) and is_prime( __magic_name__ ), "'ans' must been a prime number and from type int" return ans def _A ( __magic_name__ , __magic_name__ ): assert ( is_prime(__magic_name__ ) and is_prime(__magic_name__ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowercase__ = p_number_a + 1 # jump to the next number lowercase__ = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__magic_name__ ): number += 1 while number < p_number_a: ans.append(__magic_name__ ) number += 1 # fetch the next prime number. while not is_prime(__magic_name__ ): number += 1 # precondition assert ( isinstance(__magic_name__ , __magic_name__ ) and ans[0] != p_number_a and ans[len(__magic_name__ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n >= 1), "'n' must been int and >= 1" lowercase__ = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(__magic_name__ ) # precondition assert ans[0] == 1 and ans[len(__magic_name__ ) - 1] == n, "Error in function getDivisiors(...)" return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and ( number > 1 ), "'number' must been an int and >= 1" lowercase__ = get_divisors(__magic_name__ ) # precondition assert ( isinstance(__magic_name__ , __magic_name__ ) and (divisors[0] == 1) and (divisors[len(__magic_name__ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def _A ( __magic_name__ , __magic_name__ ): assert ( isinstance(__magic_name__ , __magic_name__ ) and isinstance(__magic_name__ , __magic_name__ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowercase__ = gcd(abs(__magic_name__ ) , abs(__magic_name__ ) ) # precondition assert ( isinstance(__magic_name__ , __magic_name__ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n >= 0), "'n' must been a int and >= 0" lowercase__ = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def _A ( __magic_name__ ): assert isinstance(__magic_name__ , __magic_name__ ) and (n >= 0), "'n' must been an int and >= 0" lowercase__ = 0 lowercase__ = 1 lowercase__ = 1 # this will be return for _ in range(n - 1 ): lowercase__ = ans ans += fiba lowercase__ = tmp return ans
655
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 # setable values __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = None @classmethod def UpperCAmelCase ( cls :Union[str, Any] , _lowercase :CommonSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray ): '''simple docstring''' return cls(common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase ) @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase ( lowercase_ , lowercase_ ): __lowerCamelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCamelCase = 42 @property def UpperCAmelCase ( self :List[str] ): '''simple docstring''' return True @register_to_config def __init__( self :str , _lowercase :int = 10_00 , _lowercase :float = 0.0001 , _lowercase :float = 0.02 , _lowercase :str = "linear" , _lowercase :Optional[jnp.ndarray] = None , _lowercase :str = "fixed_small" , _lowercase :bool = True , _lowercase :str = "epsilon" , _lowercase :jnp.dtype = jnp.floataa , ): '''simple docstring''' lowercase__ = dtype def UpperCAmelCase ( self :str , _lowercase :Optional[CommonSchedulerState] = None ): '''simple docstring''' if common is None: lowercase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowercase__ = jnp.array(1.0 , dtype=self.dtype ) lowercase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowercase , init_noise_sigma=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase ( self :List[str] , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :Tuple = () ): '''simple docstring''' lowercase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowercase__ = (jnp.arange(0 , _lowercase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowercase , timesteps=_lowercase , ) def UpperCAmelCase ( self :Tuple , _lowercase :DDPMSchedulerState , _lowercase :int , _lowercase :List[str]=None , _lowercase :Tuple=None ): '''simple docstring''' lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowercase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowercase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowercase__ = jnp.clip(_lowercase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowercase__ = jnp.log(jnp.clip(_lowercase , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowercase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowercase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowercase__ = variance lowercase__ = state.common.betas[t] lowercase__ = (predicted_variance + 1) / 2 lowercase__ = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase ( self :Optional[int] , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :int , _lowercase :jnp.ndarray , _lowercase :Optional[jax.random.KeyArray] = None , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = timestep if key is None: lowercase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ = jnp.split(_lowercase , sample.shape[1] , axis=1 ) else: lowercase__ = None # 1. compute alphas, betas lowercase__ = state.common.alphas_cumprod[t] lowercase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowercase__ = 1 - alpha_prod_t lowercase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowercase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowercase__ = model_output elif self.config.prediction_type == "v_prediction": lowercase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowercase__ = jnp.clip(_lowercase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowercase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowercase__ = jax.random.split(_lowercase , num=1 ) lowercase__ = jax.random.normal(_lowercase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowercase , _lowercase , predicted_variance=_lowercase ) ** 0.5) * noise lowercase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowercase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowercase , state=_lowercase ) def UpperCAmelCase ( self :int , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return add_noise_common(state.common , _lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :Dict , _lowercase :DDPMSchedulerState , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , _lowercase :jnp.ndarray , ): '''simple docstring''' return get_velocity_common(state.common , _lowercase , _lowercase , _lowercase ) def __len__( self :List[str] ): '''simple docstring''' return self.config.num_train_timesteps
655
1
def _A ( __magic_name__ , __magic_name__ ): return price * (1 + tax_rate) if __name__ == "__main__": print(F"""{price_plus_tax(100, 0.25) = }""") print(F"""{price_plus_tax(1_25.50, 0.05) = }""")
655
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _snake_case = logging.get_logger(__name__) _snake_case = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _snake_case = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCAmelCase : __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Model type selected in the list: ' + ', '.join(lowercase_ )} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) __lowerCamelCase = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) __lowerCamelCase = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) __lowerCamelCase = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) __lowerCamelCase = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'train' __lowerCamelCase = 'dev' class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 def __init__( self :Optional[Any] , _lowercase :SquadDataTrainingArguments , _lowercase :PreTrainedTokenizer , _lowercase :Optional[int] = None , _lowercase :Union[str, Split] = Split.train , _lowercase :Optional[bool] = False , _lowercase :Optional[str] = None , _lowercase :Optional[str] = "pt" , ): '''simple docstring''' lowercase__ = args lowercase__ = is_language_sensitive lowercase__ = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_lowercase , _lowercase ): try: lowercase__ = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) lowercase__ = mode # Load data features from cache or dataset file lowercase__ = "v2" if args.version_2_with_negative else "v1" lowercase__ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase__ = cached_features_file + ".lock" with FileLock(_lowercase ): if os.path.exists(_lowercase ) and not args.overwrite_cache: lowercase__ = time.time() lowercase__ = torch.load(_lowercase ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowercase__ = self.old_features["features"] lowercase__ = self.old_features.get("dataset" , _lowercase ) lowercase__ = self.old_features.get("examples" , _lowercase ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: lowercase__ = self.processor.get_dev_examples(args.data_dir ) else: lowercase__ = self.processor.get_train_examples(args.data_dir ) lowercase__ , lowercase__ = squad_convert_examples_to_features( examples=self.examples , tokenizer=_lowercase , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_lowercase , ) lowercase__ = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _lowercase , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self :Dict ): '''simple docstring''' return len(self.features ) def __getitem__( self :Any , _lowercase :Any ): '''simple docstring''' lowercase__ = self.features[i] lowercase__ = torch.tensor(feature.input_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.attention_mask , dtype=torch.long ) lowercase__ = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowercase__ = torch.tensor(feature.cls_index , dtype=torch.long ) lowercase__ = torch.tensor(feature.p_mask , dtype=torch.float ) lowercase__ = torch.tensor(feature.is_impossible , dtype=torch.float ) lowercase__ = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowercase__ = torch.tensor(feature.start_position , dtype=torch.long ) lowercase__ = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
655
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _snake_case = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ _snake_case = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ _snake_case = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _A ( __magic_name__ , __magic_name__ ): return float((preds == labels).mean() ) def _A ( __magic_name__ , __magic_name__ , __magic_name__="binary" ): lowercase__ = simple_accuracy(__magic_name__ , __magic_name__ ) lowercase__ = float(fa_score(y_true=__magic_name__ , y_pred=__magic_name__ , average=__magic_name__ ) ) return { "accuracy": acc, "f1": fa, } def _A ( __magic_name__ , __magic_name__ ): lowercase__ = {} for id_pred, label in zip(__magic_name__ , __magic_name__ ): lowercase__ = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' lowercase__ = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowercase__ = [(pred, label)] lowercase__ , lowercase__ = [], [] for question, preds_labels in question_map.items(): lowercase__ , lowercase__ = zip(*__magic_name__ ) lowercase__ = fa_score(y_true=__magic_name__ , y_pred=__magic_name__ , average="macro" ) fas.append(__magic_name__ ) lowercase__ = int(sum(pred == label for pred, label in preds_labels ) == len(__magic_name__ ) ) ems.append(__magic_name__ ) lowercase__ = float(sum(__magic_name__ ) / len(__magic_name__ ) ) lowercase__ = sum(__magic_name__ ) / len(__magic_name__ ) lowercase__ = float(fa_score(y_true=__magic_name__ , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def UpperCAmelCase ( self :str ): '''simple docstring''' if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def UpperCAmelCase ( self :Union[str, Any] , _lowercase :List[Any] , _lowercase :Any ): '''simple docstring''' if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(_lowercase , _lowercase )} elif self.config_name == "cb": return acc_and_fa(_lowercase , _lowercase , fa_avg="macro" ) elif self.config_name == "record": lowercase__ = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] lowercase__ = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(_lowercase , _lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(_lowercase , _lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(_lowercase , _lowercase )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
655
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = """▁""" _snake_case = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} _snake_case = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } _snake_case = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } _snake_case = { """ernie-m-base""": 514, """ernie-m-large""": 514, } _snake_case = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ["input_ids"] __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = RESOURCE_FILES_NAMES def __init__( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=False , _lowercase :Dict="utf8" , _lowercase :Optional[Any]="[UNK]" , _lowercase :Optional[int]="[SEP]" , _lowercase :List[str]="[PAD]" , _lowercase :Dict="[CLS]" , _lowercase :Optional[Any]="[MASK]" , _lowercase :Optional[Dict[str, Any]] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , vocab_file=_lowercase , encoding=_lowercase , sp_model_kwargs=self.sp_model_kwargs , **_lowercase , ) lowercase__ = do_lower_case lowercase__ = sentencepiece_model_ckpt lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase__ = self.load_vocab(filepath=_lowercase ) else: lowercase__ = {self.sp_model.id_to_piece(_lowercase ): id for id in range(self.sp_model.get_piece_size() )} lowercase__ = {v: k for k, v in self.vocab.items()} def UpperCAmelCase ( self :Any , _lowercase :Dict ): '''simple docstring''' if text is None: return None lowercase__ = self.tokenize(_lowercase ) lowercase__ , lowercase__ = "", [] for i, ch in enumerate(_lowercase ): if ch in self.SP_CHAR_MAPPING: lowercase__ = self.SP_CHAR_MAPPING.get(_lowercase ) else: lowercase__ = unicodedata.normalize("NFKC" , _lowercase ) if self.is_whitespace(_lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(_lowercase ) ) lowercase__ , lowercase__ , lowercase__ = normalized_text, [], 0 if self.do_lower_case: lowercase__ = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase__ = token[1:] lowercase__ = text[offset:].index(_lowercase ) + offset lowercase__ = start + len(_lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase__ = end return token_mapping @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return len(self.vocab ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self :Any ): '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self :Optional[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCAmelCase ( self :List[str] , _lowercase :Optional[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(_lowercase , _lowercase ) for c in text) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any]=False , _lowercase :Optional[int]=64 , _lowercase :Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get("enable_sampling" ) is True: lowercase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: lowercase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: lowercase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: lowercase__ = self.sp_model.EncodeAsPieces(_lowercase ) else: lowercase__ = self.sp_model.SampleEncodeAsPieces(_lowercase , _lowercase , _lowercase ) lowercase__ = [] for pi, piece in enumerate(_lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(_lowercase ) and pi != 0: new_pieces.append(_lowercase ) continue else: continue lowercase__ = 0 for i, chunk in enumerate(_lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(_lowercase ) or self.is_punct(_lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(_lowercase ) lowercase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i if len(_lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCAmelCase ( self :Tuple , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = self.convert_ids_to_tokens(_lowercase ) lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' return self.vocab.get(_lowercase , self.vocab.get(self.unk_token ) ) def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' return self.reverse_vocab.get(_lowercase , self.unk_token ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :Tuple=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCAmelCase ( self :Dict , _lowercase :int , _lowercase :Union[str, Any]=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Dict=None , _lowercase :Optional[Any]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(_lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(_lowercase ) + 1) + [1] * (len(_lowercase ) + 3) def UpperCAmelCase ( self :str , _lowercase :Optional[int] ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCAmelCase ( self :int , _lowercase :Dict ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCAmelCase ( self :List[str] , _lowercase :List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(_lowercase ) == 1: lowercase__ = unicodedata.category(_lowercase ) if cat == "Zs": return True return False def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = {} with io.open(_lowercase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(_lowercase ): lowercase__ = line.rstrip("\n" ) lowercase__ = int(_lowercase ) return token_to_idx def UpperCAmelCase ( self :List[str] , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' lowercase__ = 0 if os.path.isdir(_lowercase ): lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowercase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(_lowercase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowercase__ = token_index writer.write(token + "\n" ) index += 1 lowercase__ = os.path.join(_lowercase , "sentencepiece.bpe.model" ) with open(_lowercase , "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(_lowercase ) return (vocab_file,)
655
1
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = LayoutLMTokenizer __lowerCamelCase = LayoutLMTokenizerFast __lowerCamelCase = True __lowerCamelCase = True def UpperCAmelCase ( self :Any ): '''simple docstring''' super().setUp() lowercase__ = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowercase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def UpperCAmelCase ( self :List[Any] , **_lowercase :Optional[Any] ): '''simple docstring''' return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowercase ) def UpperCAmelCase ( self :str , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "UNwant\u00E9d,running" lowercase__ = "unwanted, running" return input_text, output_text def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.tokenizer_class(self.vocab_file ) lowercase__ = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(_lowercase , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ) , [7, 4, 5, 10, 8, 9] ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' pass
655
def _A ( __magic_name__ ): lowercase__ = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def _A ( __magic_name__ = 100 ): lowercase__ = 1 lowercase__ = 2 for i in range(2 , max_n + 1 ): lowercase__ = pre_numerator lowercase__ = 2 * i // 3 if i % 3 == 0 else 1 lowercase__ = cur_numerator lowercase__ = e_cont * pre_numerator + temp return sum_digits(__magic_name__ ) if __name__ == "__main__": print(F"""{solution() = }""")
655
1
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class lowerCAmelCase : def __init__( self :List[Any] , _lowercase :str , _lowercase :Tuple=2 , _lowercase :Dict=32 , _lowercase :Optional[Any]=16 , _lowercase :int=3 , _lowercase :Optional[int]=True , _lowercase :Union[str, Any]=True , _lowercase :Optional[int]=32 , _lowercase :List[str]=4 , _lowercase :Optional[Any]=[0, 1, 2, 3] , _lowercase :Optional[Any]=4 , _lowercase :int=37 , _lowercase :Dict="gelu" , _lowercase :List[str]=0.1 , _lowercase :int=0.1 , _lowercase :List[Any]=0.02 , _lowercase :Any=3 , _lowercase :Any=[1, 3_84, 24, 24] , _lowercase :Any=True , _lowercase :str=None , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = is_training lowercase__ = use_labels lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = backbone_out_indices lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = initializer_range lowercase__ = num_labels lowercase__ = backbone_featmap_shape lowercase__ = scope lowercase__ = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) lowercase__ = (image_size // patch_size) ** 2 lowercase__ = num_patches + 1 def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowercase__ = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [96, 1_92, 3_84, 7_68], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=_lowercase , backbone_featmap_shape=self.backbone_featmap_shape , ) def UpperCAmelCase ( self :int , _lowercase :List[str] , _lowercase :Optional[int] , _lowercase :List[str] ): '''simple docstring''' lowercase__ = DPTModel(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self :Dict , _lowercase :Optional[int] , _lowercase :Optional[Any] , _lowercase :int ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = DPTForDepthEstimation(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :str , _lowercase :Any , _lowercase :Union[str, Any] ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = DPTForSemanticSegmentation(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , labels=_lowercase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () __lowerCamelCase = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = DPTModelTester(self ) lowercase__ = ConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def UpperCAmelCase ( self :int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="DPT does not use inputs_embeds" ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' pass def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowercase , nn.Linear ) ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*_lowercase ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = True if model_class in get_values(_lowercase ): continue lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.train() lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) lowercase__ = model(**_lowercase ).loss loss.backward() def UpperCAmelCase ( self :Any ): '''simple docstring''' for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = False lowercase__ = True if model_class in get_values(_lowercase ) or not model_class.supports_gradient_checkpointing: continue lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.gradient_checkpointing_enable() model.train() lowercase__ = self._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) lowercase__ = model(**_lowercase ).loss loss.backward() def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = _config_zero_init(_lowercase ) for model_class in self.all_model_classes: lowercase__ = model_class(config=_lowercase ) # Skip the check for the backbone lowercase__ = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": lowercase__ = [f'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' pass @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: lowercase__ = DPTModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = "add" with self.assertRaises(_lowercase ): lowercase__ = DPTForDepthEstimation(_lowercase ) def _A ( ): lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision @slow class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = DPTImageProcessor.from_pretrained("Intel/dpt-hybrid-midas" ) lowercase__ = DPTForDepthEstimation.from_pretrained("Intel/dpt-hybrid-midas" ).to(_lowercase ) lowercase__ = prepare_img() lowercase__ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): lowercase__ = model(**_lowercase ) lowercase__ = outputs.predicted_depth # verify the predicted depth lowercase__ = torch.Size((1, 3_84, 3_84) ) self.assertEqual(predicted_depth.shape , _lowercase ) lowercase__ = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 1_00 , _lowercase , atol=1e-4 ) )
655
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case = logging.get_logger(__name__) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'AutoTokenizer' __lowerCamelCase = ['tokenizer'] __lowerCamelCase = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self :Dict , _lowercase :List[str] , _lowercase :List[Any]=None ): '''simple docstring''' super().__init__(_lowercase ) lowercase__ = speaker_embeddings @classmethod def UpperCAmelCase ( cls :Any , _lowercase :int , _lowercase :str="speaker_embeddings_path.json" , **_lowercase :List[str] ): '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase__ = get_file_from_repo( _lowercase , _lowercase , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(_lowercase , _lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase__ = None else: with open(_lowercase ) as speaker_embeddings_json: lowercase__ = json.load(_lowercase ) else: lowercase__ = None lowercase__ = AutoTokenizer.from_pretrained(_lowercase , **_lowercase ) return cls(tokenizer=_lowercase , speaker_embeddings=_lowercase ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :List[str]="speaker_embeddings_path.json" , _lowercase :Any="speaker_embeddings" , _lowercase :bool = False , **_lowercase :Any , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(_lowercase , _lowercase , "v2" ) , exist_ok=_lowercase ) lowercase__ = {} lowercase__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase__ = self._load_voice_preset(_lowercase ) lowercase__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , _lowercase , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_lowercase , ) lowercase__ = os.path.join(_lowercase , f'''{prompt_key}_{key}.npy''' ) lowercase__ = tmp_dict with open(os.path.join(_lowercase , _lowercase ) , "w" ) as fp: json.dump(_lowercase , _lowercase ) super().save_pretrained(_lowercase , _lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :str = None , **_lowercase :List[Any] ): '''simple docstring''' lowercase__ = self.speaker_embeddings[voice_preset] lowercase__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase__ = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , _lowercase ) , cache_dir=kwargs.pop("cache_dir" , _lowercase ) , force_download=kwargs.pop("force_download" , _lowercase ) , proxies=kwargs.pop("proxies" , _lowercase ) , resume_download=kwargs.pop("resume_download" , _lowercase ) , local_files_only=kwargs.pop("local_files_only" , _lowercase ) , use_auth_token=kwargs.pop("use_auth_token" , _lowercase ) , revision=kwargs.pop("revision" , _lowercase ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) lowercase__ = np.load(_lowercase ) return voice_preset_dict def UpperCAmelCase ( self :Optional[int] , _lowercase :Optional[dict] = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Optional[Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=None , _lowercase :List[str]="pt" , _lowercase :List[Any]=2_56 , _lowercase :List[str]=False , _lowercase :Union[str, Any]=True , _lowercase :Dict=False , **_lowercase :Tuple , ): '''simple docstring''' if voice_preset is not None and not isinstance(_lowercase , _lowercase ): if ( isinstance(_lowercase , _lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase__ = self._load_voice_preset(_lowercase ) else: if isinstance(_lowercase , _lowercase ) and not voice_preset.endswith(".npz" ): lowercase__ = voice_preset + ".npz" lowercase__ = np.load(_lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(_lowercase , **_lowercase ) lowercase__ = BatchFeature(data=_lowercase , tensor_type=_lowercase ) lowercase__ = self.tokenizer( _lowercase , return_tensors=_lowercase , padding="max_length" , max_length=_lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , add_special_tokens=_lowercase , **_lowercase , ) if voice_preset is not None: lowercase__ = voice_preset return encoded_text
655
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def _A ( __magic_name__ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(__magic_name__ , __magic_name__ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(__magic_name__ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=__magic_name__ , num_labels=1000 , idalabel=__magic_name__ , labelaid=__magic_name__ , ) return config def _A ( __magic_name__ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def _A ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) return im @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__=False ): lowercase__ = get_config(__magic_name__ ) # load original model from timm lowercase__ = create_model(__magic_name__ , pretrained=__magic_name__ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(__magic_name__ ) model.eval() model.load_state_dict(__magic_name__ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=__magic_name__ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=__magic_name__ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=__magic_name__ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=__magic_name__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(__magic_name__ ).unsqueeze(0 ) lowercase__ = processor(__magic_name__ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(__magic_name__ , __magic_name__ ) # verify logits with torch.no_grad(): lowercase__ = model(__magic_name__ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(__magic_name__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__magic_name__ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) processor.save_pretrained(__magic_name__ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) _snake_case = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
655
import math import random def _A ( __magic_name__ , __magic_name__ = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value _snake_case = 0.02 def _A ( __magic_name__ , __magic_name__ ): lowercase__ = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__magic_name__ ): # Forward propagation lowercase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? lowercase__ = (expected / 100) - layer_a # Error delta lowercase__ = layer_1_error * sigmoid_function(__magic_name__ , __magic_name__ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input("""Expected value: """)) _snake_case = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
655
1
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = """▁""" _snake_case = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} _snake_case = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } _snake_case = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } _snake_case = { """ernie-m-base""": 514, """ernie-m-large""": 514, } _snake_case = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = ["input_ids"] __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = RESOURCE_FILES_NAMES def __init__( self :Union[str, Any] , _lowercase :Union[str, Any] , _lowercase :Optional[Any]=None , _lowercase :List[str]=False , _lowercase :Dict="utf8" , _lowercase :Optional[Any]="[UNK]" , _lowercase :Optional[int]="[SEP]" , _lowercase :List[str]="[PAD]" , _lowercase :Dict="[CLS]" , _lowercase :Optional[Any]="[MASK]" , _lowercase :Optional[Dict[str, Any]] = None , **_lowercase :Tuple , ): '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , vocab_file=_lowercase , encoding=_lowercase , sp_model_kwargs=self.sp_model_kwargs , **_lowercase , ) lowercase__ = do_lower_case lowercase__ = sentencepiece_model_ckpt lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase__ = self.load_vocab(filepath=_lowercase ) else: lowercase__ = {self.sp_model.id_to_piece(_lowercase ): id for id in range(self.sp_model.get_piece_size() )} lowercase__ = {v: k for k, v in self.vocab.items()} def UpperCAmelCase ( self :Any , _lowercase :Dict ): '''simple docstring''' if text is None: return None lowercase__ = self.tokenize(_lowercase ) lowercase__ , lowercase__ = "", [] for i, ch in enumerate(_lowercase ): if ch in self.SP_CHAR_MAPPING: lowercase__ = self.SP_CHAR_MAPPING.get(_lowercase ) else: lowercase__ = unicodedata.normalize("NFKC" , _lowercase ) if self.is_whitespace(_lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(_lowercase ) ) lowercase__ , lowercase__ , lowercase__ = normalized_text, [], 0 if self.do_lower_case: lowercase__ = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase__ = token[1:] lowercase__ = text[offset:].index(_lowercase ) + offset lowercase__ = start + len(_lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase__ = end return token_mapping @property def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return len(self.vocab ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self :Any ): '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self :Optional[Any] , _lowercase :Dict ): '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCAmelCase ( self :List[str] , _lowercase :Optional[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(_lowercase , _lowercase ) for c in text) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any]=False , _lowercase :Optional[int]=64 , _lowercase :Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get("enable_sampling" ) is True: lowercase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: lowercase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: lowercase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: lowercase__ = self.sp_model.EncodeAsPieces(_lowercase ) else: lowercase__ = self.sp_model.SampleEncodeAsPieces(_lowercase , _lowercase , _lowercase ) lowercase__ = [] for pi, piece in enumerate(_lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(_lowercase ) and pi != 0: new_pieces.append(_lowercase ) continue else: continue lowercase__ = 0 for i, chunk in enumerate(_lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(_lowercase ) or self.is_punct(_lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(_lowercase ) lowercase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase__ = i if len(_lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCAmelCase ( self :Tuple , _lowercase :Tuple ): '''simple docstring''' lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Any , _lowercase :str ): '''simple docstring''' lowercase__ = self.convert_ids_to_tokens(_lowercase ) lowercase__ = "".join(_lowercase ).replace(_lowercase , " " ).strip() return out_string def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' return self.vocab.get(_lowercase , self.vocab.get(self.unk_token ) ) def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' return self.reverse_vocab.get(_lowercase , self.unk_token ) def UpperCAmelCase ( self :Any , _lowercase :Any , _lowercase :Tuple=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCAmelCase ( self :Dict , _lowercase :int , _lowercase :Union[str, Any]=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Dict=None , _lowercase :Optional[Any]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] def UpperCAmelCase ( self :int , _lowercase :List[int] , _lowercase :Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(_lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(_lowercase ) + 1) + [1] * (len(_lowercase ) + 3) def UpperCAmelCase ( self :str , _lowercase :Optional[int] ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def UpperCAmelCase ( self :Tuple , _lowercase :List[str] ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCAmelCase ( self :int , _lowercase :Dict ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCAmelCase ( self :List[str] , _lowercase :List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(_lowercase ) == 1: lowercase__ = unicodedata.category(_lowercase ) if cat == "Zs": return True return False def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = {} with io.open(_lowercase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(_lowercase ): lowercase__ = line.rstrip("\n" ) lowercase__ = int(_lowercase ) return token_to_idx def UpperCAmelCase ( self :List[str] , _lowercase :str , _lowercase :Optional[str] = None ): '''simple docstring''' lowercase__ = 0 if os.path.isdir(_lowercase ): lowercase__ = os.path.join( _lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: lowercase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(_lowercase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' " Please check that the vocabulary is not corrupted!" ) lowercase__ = token_index writer.write(token + "\n" ) index += 1 lowercase__ = os.path.join(_lowercase , "sentencepiece.bpe.model" ) with open(_lowercase , "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(_lowercase ) return (vocab_file,)
655
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'van' def __init__( self :Optional[Any] , _lowercase :Dict=2_24 , _lowercase :Union[str, Any]=3 , _lowercase :List[Any]=[7, 3, 3, 3] , _lowercase :Any=[4, 2, 2, 2] , _lowercase :Union[str, Any]=[64, 1_28, 3_20, 5_12] , _lowercase :List[Any]=[3, 3, 12, 3] , _lowercase :Dict=[8, 8, 4, 4] , _lowercase :int="gelu" , _lowercase :List[Any]=0.02 , _lowercase :List[Any]=1e-6 , _lowercase :Any=1e-2 , _lowercase :int=0.0 , _lowercase :int=0.0 , **_lowercase :Dict , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = image_size lowercase__ = num_channels lowercase__ = patch_sizes lowercase__ = strides lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = mlp_ratios lowercase__ = hidden_act lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = layer_scale_init_value lowercase__ = drop_path_rate lowercase__ = dropout_rate
655
1
from __future__ import annotations def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = position lowercase__ = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowercase__ = [] for position in positions: lowercase__ , lowercase__ = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(__magic_name__ ) return permissible_positions def _A ( __magic_name__ ): return not any(elem == 0 for row in board for elem in row ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if is_complete(__magic_name__ ): return True for position in get_valid_pos(__magic_name__ , len(__magic_name__ ) ): lowercase__ , lowercase__ = position if board[y][x] == 0: lowercase__ = curr + 1 if open_knight_tour_helper(__magic_name__ , __magic_name__ , curr + 1 ): return True lowercase__ = 0 return False def _A ( __magic_name__ ): lowercase__ = [[0 for i in range(__magic_name__ )] for j in range(__magic_name__ )] for i in range(__magic_name__ ): for j in range(__magic_name__ ): lowercase__ = 1 if open_knight_tour_helper(__magic_name__ , (i, j) , 1 ): return board lowercase__ = 0 lowercase__ = f'''Open Kight Tour cannot be performed on a board of size {n}''' raise ValueError(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCAmelCase ( enum.Enum ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 @add_end_docstrings(lowercase_ ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self :Any , *_lowercase :Optional[Any] , **_lowercase :Union[str, Any] ): '''simple docstring''' super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowercase__ = None if self.model.config.prefix is not None: lowercase__ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowercase__ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowercase__ , lowercase__ , lowercase__ = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) lowercase__ = {**self._preprocess_params, **preprocess_params} lowercase__ = {**self._forward_params, **forward_params} def UpperCAmelCase ( self :Tuple , _lowercase :Optional[Any]=None , _lowercase :List[Any]=None , _lowercase :List[str]=None , _lowercase :Optional[Any]=None , _lowercase :Optional[int]=None , _lowercase :Any=None , _lowercase :Any=None , _lowercase :Dict=None , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ = {} if prefix is not None: lowercase__ = prefix if prefix: lowercase__ = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) lowercase__ = handle_long_generation preprocess_params.update(_lowercase ) lowercase__ = generate_kwargs lowercase__ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) lowercase__ = ReturnType.TENSORS if return_type is not None: lowercase__ = return_type if clean_up_tokenization_spaces is not None: lowercase__ = clean_up_tokenization_spaces if stop_sequence is not None: lowercase__ = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) lowercase__ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase ( self :int , *_lowercase :Optional[int] , **_lowercase :List[str] ): '''simple docstring''' if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self :Union[str, Any] , _lowercase :Dict , **_lowercase :Tuple ): '''simple docstring''' return super().__call__(_lowercase , **_lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple , _lowercase :Optional[int]="" , _lowercase :Tuple=None , **_lowercase :List[str] ): '''simple docstring''' lowercase__ = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) lowercase__ = prompt_text if handle_long_generation == "hole": lowercase__ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: lowercase__ = generate_kwargs["max_new_tokens"] else: lowercase__ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: lowercase__ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) lowercase__ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: lowercase__ = inputs["attention_mask"][:, -keep_length:] return inputs def UpperCAmelCase ( self :str , _lowercase :int , **_lowercase :str ): '''simple docstring''' lowercase__ = model_inputs["input_ids"] lowercase__ = model_inputs.get("attention_mask" , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: lowercase__ = None lowercase__ = None lowercase__ = 1 else: lowercase__ = input_ids.shape[0] lowercase__ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowercase__ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: lowercase__ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: lowercase__ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowercase__ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowercase__ = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) lowercase__ = generated_sequence.shape[0] if self.framework == "pt": lowercase__ = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": lowercase__ = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCAmelCase ( self :Any , _lowercase :Tuple , _lowercase :str=ReturnType.FULL_TEXT , _lowercase :Dict=True ): '''simple docstring''' lowercase__ = model_outputs["generated_sequence"][0] lowercase__ = model_outputs["input_ids"] lowercase__ = model_outputs["prompt_text"] lowercase__ = generated_sequence.numpy().tolist() lowercase__ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowercase__ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowercase__ = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowercase__ = 0 else: lowercase__ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: lowercase__ = prompt_text + text[prompt_length:] else: lowercase__ = text[prompt_length:] lowercase__ = {"generated_text": all_text} records.append(_lowercase ) return records
655
1
_snake_case = """Alexander Joslin""" import operator as op from .stack import Stack def _A ( __magic_name__ ): lowercase__ = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} lowercase__ = Stack() lowercase__ = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__magic_name__ ) ) elif i in operators: # RULE 2 operator_stack.push(__magic_name__ ) elif i == ")": # RULE 4 lowercase__ = operator_stack.peek() operator_stack.pop() lowercase__ = operand_stack.peek() operand_stack.pop() lowercase__ = operand_stack.peek() operand_stack.pop() lowercase__ = operators[opr](__magic_name__ , __magic_name__ ) operand_stack.push(__magic_name__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": _snake_case = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
655
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _snake_case = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _snake_case = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def _A ( __magic_name__ ): lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__magic_name__ )[0] @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) lowercase__ = data.reshape(__magic_name__ , __magic_name__ , __magic_name__ , 1 ) return data @deprecated(__magic_name__ , "Please use tf.one_hot on tensors." ) def _A ( __magic_name__ , __magic_name__ ): lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__magic_name__ ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__magic_name__ , "Please use tf.data to implement this functionality." ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=10 ): print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__magic_name__ ) as bytestream: lowercase__ = _readaa(__magic_name__ ) if magic != 2049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowercase__ = _readaa(__magic_name__ ) lowercase__ = bytestream.read(__magic_name__ ) lowercase__ = numpy.frombuffer(__magic_name__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__magic_name__ , __magic_name__ ) return labels class lowerCAmelCase : @deprecated( _lowercase , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :Tuple=False , _lowercase :str=False , _lowercase :Dict=dtypes.floataa , _lowercase :Optional[Any]=True , _lowercase :Any=None , ): '''simple docstring''' lowercase__ , lowercase__ = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowercase__ = 1_00_00 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'''images.shape: {images.shape} labels.shape: {labels.shape}''' lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(_lowercase , 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._images @property def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' return self._labels @property def UpperCAmelCase ( self :Dict ): '''simple docstring''' return self._num_examples @property def UpperCAmelCase ( self :Tuple ): '''simple docstring''' return self._epochs_completed def UpperCAmelCase ( self :str , _lowercase :Union[str, Any] , _lowercase :Any=False , _lowercase :Union[str, Any]=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 7_84 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__magic_name__ , "Please write your own downloading logic." ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): if not gfile.Exists(__magic_name__ ): gfile.MakeDirs(__magic_name__ ) lowercase__ = os.path.join(__magic_name__ , __magic_name__ ) if not gfile.Exists(__magic_name__ ): urllib.request.urlretrieve(__magic_name__ , __magic_name__ ) # noqa: S310 with gfile.GFile(__magic_name__ ) as f: lowercase__ = f.size() print("Successfully downloaded" , __magic_name__ , __magic_name__ , "bytes." ) return filepath @deprecated( __magic_name__ , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def _A ( __magic_name__ , __magic_name__=False , __magic_name__=False , __magic_name__=dtypes.floataa , __magic_name__=True , __magic_name__=5000 , __magic_name__=None , __magic_name__=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__magic_name__ , one_hot=__magic_name__ , dtype=__magic_name__ , seed=__magic_name__ ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = "train-images-idx3-ubyte.gz" lowercase__ = "train-labels-idx1-ubyte.gz" lowercase__ = "t10k-images-idx3-ubyte.gz" lowercase__ = "t10k-labels-idx1-ubyte.gz" lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + train_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_images_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_images(__magic_name__ ) lowercase__ = _maybe_download( __magic_name__ , __magic_name__ , source_url + test_labels_file ) with gfile.Open(__magic_name__ , "rb" ) as f: lowercase__ = _extract_labels(__magic_name__ , one_hot=__magic_name__ ) if not 0 <= validation_size <= len(__magic_name__ ): lowercase__ = ( "Validation size should be between 0 and " f'''{len(__magic_name__ )}. Received: {validation_size}.''' ) raise ValueError(__magic_name__ ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {"dtype": dtype, "reshape": reshape, "seed": seed} lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) lowercase__ = _DataSet(__magic_name__ , __magic_name__ , **__magic_name__ ) return _Datasets(train=__magic_name__ , validation=__magic_name__ , test=__magic_name__ )
655
1
def _A ( __magic_name__ , __magic_name__ ): lowercase__ = [[] for _ in range(__magic_name__ )] lowercase__ = key - 1 if key <= 0: raise ValueError("Height of grid can't be 0 or negative" ) if key == 1 or len(__magic_name__ ) <= key: return input_string for position, character in enumerate(__magic_name__ ): lowercase__ = position % (lowest * 2) # puts it in bounds lowercase__ = min(__magic_name__ , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(__magic_name__ ) lowercase__ = ["".join(__magic_name__ ) for row in temp_grid] lowercase__ = "".join(__magic_name__ ) return output_string def _A ( __magic_name__ , __magic_name__ ): lowercase__ = [] lowercase__ = key - 1 if key <= 0: raise ValueError("Height of grid can't be 0 or negative" ) if key == 1: return input_string lowercase__ = [[] for _ in range(__magic_name__ )] # generates template for position in range(len(__magic_name__ ) ): lowercase__ = position % (lowest * 2) # puts it in bounds lowercase__ = min(__magic_name__ , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("*" ) lowercase__ = 0 for row in temp_grid: # fills in the characters lowercase__ = input_string[counter : counter + len(__magic_name__ )] grid.append(list(__magic_name__ ) ) counter += len(__magic_name__ ) lowercase__ = "" # reads as zigzag for position in range(len(__magic_name__ ) ): lowercase__ = position % (lowest * 2) # puts it in bounds lowercase__ = min(__magic_name__ , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def _A ( __magic_name__ ): lowercase__ = {} for key_guess in range(1 , len(__magic_name__ ) ): # tries every key lowercase__ = decrypt(__magic_name__ , __magic_name__ ) return results if __name__ == "__main__": import doctest doctest.testmod()
655
from __future__ import annotations class lowerCAmelCase : def __init__( self :Union[str, Any] , _lowercase :List[Any]=None ): '''simple docstring''' lowercase__ = data lowercase__ = None def __repr__( self :Dict ): '''simple docstring''' lowercase__ = [] lowercase__ = self while temp: string_rep.append(f'''{temp.data}''' ) lowercase__ = temp.next return "->".join(_lowercase ) def _A ( __magic_name__ ): if not elements_list: raise Exception("The Elements List is empty" ) lowercase__ = lowercase__ = Node(elements_list[0] ) for i in range(1 , len(__magic_name__ ) ): lowercase__ = Node(elements_list[i] ) lowercase__ = current.next return head def _A ( __magic_name__ ): if head_node is not None and isinstance(__magic_name__ , __magic_name__ ): print_reverse(head_node.next ) print(head_node.data ) def _A ( ): from doctest import testmod testmod() lowercase__ = make_linked_list([14, 52, 14, 12, 43] ) print("Linked List:" ) print(__magic_name__ ) print("Elements in Reverse:" ) print_reverse(__magic_name__ ) if __name__ == "__main__": main()
655
1
import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase ( lowercase_ , lowercase_ ): @register_to_config def __init__( self :Dict , *, _lowercase :int = 4 , _lowercase :int = 7_68 , _lowercase :int , _lowercase :Optional[int] , ): '''simple docstring''' super().__init__() lowercase__ = nn.Parameter(torch.zeros(_lowercase ) ) # parameters for additional clip time embeddings lowercase__ = nn.Linear(_lowercase , _lowercase ) lowercase__ = nn.Linear(_lowercase , _lowercase ) # parameters for encoder hidden states lowercase__ = clip_extra_context_tokens lowercase__ = nn.Linear( _lowercase , self.clip_extra_context_tokens * cross_attention_dim ) lowercase__ = nn.Linear(_lowercase , _lowercase ) lowercase__ = nn.LayerNorm(_lowercase ) def UpperCAmelCase ( self :List[str] , *, _lowercase :Union[str, Any] , _lowercase :Optional[Any] , _lowercase :Optional[Any] , _lowercase :Dict ): '''simple docstring''' if do_classifier_free_guidance: # Add the classifier free guidance embeddings to the image embeddings lowercase__ = image_embeddings.shape[0] lowercase__ = self.learned_classifier_free_guidance_embeddings.unsqueeze(0 ) lowercase__ = classifier_free_guidance_embeddings.expand( _lowercase , -1 ) lowercase__ = torch.cat([classifier_free_guidance_embeddings, image_embeddings] , dim=0 ) # The image embeddings batch size and the text embeddings batch size are equal assert image_embeddings.shape[0] == prompt_embeds.shape[0] lowercase__ = prompt_embeds.shape[0] # "Specifically, we modify the architecture described in Nichol et al. (2021) by projecting and # adding CLIP embeddings to the existing timestep embedding, ... lowercase__ = self.embedding_proj(_lowercase ) lowercase__ = self.clip_image_embeddings_project_to_time_embeddings(_lowercase ) lowercase__ = time_projected_image_embeddings + time_projected_prompt_embeds # ... and by projecting CLIP embeddings into four # extra tokens of context that are concatenated to the sequence of outputs from the GLIDE text encoder" lowercase__ = self.clip_extra_context_tokens_proj(_lowercase ) lowercase__ = clip_extra_context_tokens.reshape(_lowercase , -1 , self.clip_extra_context_tokens ) lowercase__ = clip_extra_context_tokens.permute(0 , 2 , 1 ) lowercase__ = self.encoder_hidden_states_proj(_lowercase ) lowercase__ = self.text_encoder_hidden_states_norm(_lowercase ) lowercase__ = torch.cat([clip_extra_context_tokens, text_encoder_hidden_states] , dim=1 ) return text_encoder_hidden_states, additive_clip_time_embeddings
655
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
1
import random from .binary_exp_mod import bin_exp_mod def _A ( __magic_name__ , __magic_name__=1000 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase__ = n - 1 lowercase__ = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase__ = 0 while count < prec: lowercase__ = random.randint(2 , n - 1 ) lowercase__ = bin_exp_mod(__magic_name__ , __magic_name__ , __magic_name__ ) if b != 1: lowercase__ = True for _ in range(__magic_name__ ): if b == n - 1: lowercase__ = False break lowercase__ = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": _snake_case = abs(int(input("""Enter bound : """).strip())) print("""Here's the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
655
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCAmelCase : def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) lowercase__ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=_lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) lowercase__ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) lowercase__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["prompt"] lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] if "image" in inputs: lowercase__ = inputs["image"] else: lowercase__ = None if "mask_image" in inputs: lowercase__ = inputs["mask_image"] else: lowercase__ = None if "original_image" in inputs: lowercase__ = inputs["original_image"] else: lowercase__ = None lowercase__ , lowercase__ = pipe.encode_prompt(_lowercase ) # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_lowercase , _lowercase , _lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_lowercase , _lowercase ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = inputs["generator"] lowercase__ = inputs["num_inference_steps"] lowercase__ = inputs["output_type"] # inputs with prompt converted to embeddings lowercase__ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: lowercase__ = image if mask_image is not None: lowercase__ = mask_image if original_image is not None: lowercase__ = original_image lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_lowercase ) lowercase__ = self.pipeline_class.from_pretrained(_lowercase ) pipe_loaded.to(_lowercase ) pipe_loaded.set_progress_bar_config(disable=_lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe_loaded(**_lowercase )[0] lowercase__ = np.abs(to_np(_lowercase ) - to_np(_lowercase ) ).max() self.assertLess(_lowercase , 1e-4 )
655
1
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def _A ( __magic_name__ ): lowercase__ = os.path.join(args.tf_model_dir , "parameters.json" ) lowercase__ = json.loads(open(__magic_name__ ).read() ) if not params: raise ValueError( f'''It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.''' ) if not args.output.endswith(".pt" ): lowercase__ = args.output + ".pt" lowercase__ = OrderedDict() with tf.device("/CPU:0" ): lowercase__ = tf.train.load_checkpoint(args.tf_model_dir ) lowercase__ = reader.get_variable_to_shape_map() for key_name in shapes.keys(): lowercase__ = reader.get_tensor(__magic_name__ ).astype(np.floataa ) if key_name.endswith("/adam_m" ) or key_name.endswith("/adam_v" ): continue if key_name.startswith("pasts/" ): if key_name.startswith("pasts/mlp" ): lowercase__ = int(key_name[9] ) elif key_name.startswith("pasts/out" ): lowercase__ = 8 lowercase__ = "model.sqout.%d.weight" % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/moe" ): lowercase__ = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/switch_gating/kernel" ): lowercase__ = "model.blocks.%d.feed_forward.mlp.router.classifier.weight" % player lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/softmlp/kernel" ): lowercase__ = "model.blocks.%d.feed_forward.soft_bypass_mlp.weight" % player lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/wo/kernel" ) or key_name.endswith("/wi/kernel" ): lowercase__ = key_name[-9:-7] for i in range(16 ): lowercase__ = "model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight" % (player, i, nlayer) lowercase__ = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/mlp" ): lowercase__ = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/p1/kernel" ): lowercase__ = "model.blocks.%d.feed_forward.mlp.wi.weight" % player lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/p1/bias" ): lowercase__ = "model.blocks.%d.feed_forward.mlp.wi.bias" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/p2/kernel" ): lowercase__ = "model.blocks.%d.feed_forward.mlp.wo.weight" % player lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/p2/bias" ): lowercase__ = "model.blocks.%d.feed_forward.mlp.wo.bias" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/ln" ): lowercase__ = int(key_name[8:].split("/" )[0] ) if key_name.endswith("/b" ): lowercase__ = "model.blocks.%d.feed_forward.norm.bias" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/g" ): lowercase__ = "model.blocks.%d.feed_forward.norm.weight" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/att" ): lowercase__ = int(key_name[9:].split("/" )[0] ) if key_name.endswith("/qkv/kernel" ): lowercase__ = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum lowercase__ = state[:, 0, :, :] lowercase__ = state[:, 1, :, :] lowercase__ = state[:, 2, :, :] lowercase__ = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase__ = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase__ = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase__ = "model.blocks.%d.self_attn.self_attn.q_proj.weight" % player lowercase__ = torch.tensor(__magic_name__ ) lowercase__ = "model.blocks.%d.self_attn.self_attn.k_proj.weight" % player lowercase__ = torch.tensor(__magic_name__ ) lowercase__ = "model.blocks.%d.self_attn.self_attn.v_proj.weight" % player lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/o/kernel" ): lowercase__ = "model.blocks.%d.self_attn.self_attn.out_proj.weight" % player lowercase__ = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/an" ): lowercase__ = int(key_name[8:].split("/" )[0] ) if key_name.endswith("/b" ): lowercase__ = "model.blocks.%d.self_attn.norm.bias" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif key_name.endswith("/g" ): lowercase__ = "model.blocks.%d.self_attn.norm.weight" % player lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) elif ( key_name.startswith("model/wte" ) or key_name.startswith("model/wpe" ) or key_name.startswith("model/ete" ) ): lowercase__ = {"wte": "embed_tokens", "wpe": "position_embeddings", "ete": "extra_position_embeddings"}[ key_name[-3:] ] lowercase__ = "model.%s.weight" % nlayer lowercase__ = vnp.copy() # same in embedded lowercase__ = torch.tensor(__magic_name__ ) if key_name.startswith("model/wte" ): lowercase__ = "lm_head.weight" lowercase__ = vnp.copy() # same in embedded lowercase__ = torch.tensor(__magic_name__ ) elif key_name.startswith("model/wob" ): lowercase__ = "final_logits_bias" lowercase__ = vnp.copy() # same in embedded lowercase__ = state.reshape((1, -1) ) lowercase__ = torch.tensor(__magic_name__ ) elif key_name == "model/dense/kernel": lowercase__ = "model.last_project.weight" lowercase__ = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase__ = torch.tensor(__magic_name__ ) elif key_name == "model/dense_1/bias": lowercase__ = "model.last_project.bias" lowercase__ = vnp.copy() # same because it is one dimensional lowercase__ = torch.tensor(__magic_name__ ) torch.save(__magic_name__ , args.output ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser( description="""model converter.""", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument("""--tf_model_dir""", metavar="""PATH""", type=str, required=True, help="""import model""") parser.add_argument("""--output""", metavar="""PATH""", type=str, required=True, help="""output model""") _snake_case = parser.parse_args() convert_tf_gptsan_to_pt(args)
655
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) lowercase__ = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase__ = model(_lowercase )["last_hidden_state"] lowercase__ = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
655
1
def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = len(__magic_name__ ) lowercase__ = [[0] * n for i in range(__magic_name__ )] for i in range(__magic_name__ ): lowercase__ = y_points[i] for i in range(2 , __magic_name__ ): for j in range(__magic_name__ , __magic_name__ ): lowercase__ = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
655
_snake_case = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def _A ( __magic_name__ ): # Make sure the supplied data is a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = f'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(__magic_name__ ) lowercase__ = "".join(bin(__magic_name__ )[2:].zfill(8 ) for byte in data ) lowercase__ = len(__magic_name__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase__ = B"=" * ((6 - len(__magic_name__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(__magic_name__ ) % 6) else: lowercase__ = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(__magic_name__ ) , 6 ) ).encode() + padding ) def _A ( __magic_name__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(__magic_name__ , __magic_name__ ) and not isinstance(__magic_name__ , __magic_name__ ): lowercase__ = ( "argument should be a bytes-like object or ASCII string, " f'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(__magic_name__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(__magic_name__ , __magic_name__ ): try: lowercase__ = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) lowercase__ = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(__magic_name__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase__ = encoded_data[:-padding] lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase__ = "".join( bin(B64_CHARSET.index(__magic_name__ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(__magic_name__ ) , 8 ) ] return bytes(__magic_name__ ) if __name__ == "__main__": import doctest doctest.testmod()
655
1
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), f'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), f'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=True ): model.train() lowercase__ = model(__magic_name__ ) lowercase__ = F.mse_loss(__magic_name__ , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(__magic_name__ ) def _A ( __magic_name__ , __magic_name__=False ): set_seed(42 ) lowercase__ = RegressionModel() lowercase__ = deepcopy(__magic_name__ ) lowercase__ = RegressionDataset(length=80 ) lowercase__ = DataLoader(__magic_name__ , batch_size=16 ) model.to(accelerator.device ) if sched: lowercase__ = AdamW(params=model.parameters() , lr=1e-3 ) lowercase__ = AdamW(params=ddp_model.parameters() , lr=1e-3 ) lowercase__ = LambdaLR(__magic_name__ , lr_lambda=lambda __magic_name__ : epoch**0.65 ) lowercase__ = LambdaLR(__magic_name__ , lr_lambda=lambda __magic_name__ : epoch**0.65 ) # Make a copy of `model` if sched: lowercase__ , lowercase__ , lowercase__ , lowercase__ = accelerator.prepare(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: lowercase__ , lowercase__ = accelerator.prepare(__magic_name__ , __magic_name__ ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _A ( __magic_name__ ): # Test when on a single CPU or GPU that the context manager does nothing lowercase__ , lowercase__ , lowercase__ = get_training_setup(__magic_name__ ) # Use a single batch lowercase__ , lowercase__ = next(iter(__magic_name__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(__magic_name__ ): step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: # Sync grads step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), f'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ = ddp_input[torch.randperm(len(__magic_name__ ) )] def _A ( __magic_name__ ): # Test on distributed setup that context manager behaves properly lowercase__ , lowercase__ , lowercase__ = get_training_setup(__magic_name__ ) # Use a single batch lowercase__ , lowercase__ = next(iter(__magic_name__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(__magic_name__ ): step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) else: # Sync grads step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ = ddp_input[torch.randperm(len(__magic_name__ ) )] def _A ( __magic_name__=False , __magic_name__=False ): lowercase__ = Accelerator( split_batches=__magic_name__ , dispatch_batches=__magic_name__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowercase__ , lowercase__ , lowercase__ = get_training_setup(__magic_name__ ) for iteration, batch in enumerate(__magic_name__ ): lowercase__ , lowercase__ = batch.values() # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Do "gradient accumulation" (noop) with accelerator.accumulate(__magic_name__ ): step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(__magic_name__ ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ = ddp_input[torch.randperm(len(__magic_name__ ) )] GradientState._reset_state() def _A ( __magic_name__=False , __magic_name__=False ): lowercase__ = Accelerator( split_batches=__magic_name__ , dispatch_batches=__magic_name__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = get_training_setup(__magic_name__ , __magic_name__ ) for iteration, batch in enumerate(__magic_name__ ): lowercase__ , lowercase__ = batch.values() # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(__magic_name__ )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(__magic_name__ ): step_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), f'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n''' lowercase__ = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(__magic_name__ )) if accelerator.num_processes > 1: check_model_parameters(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _A ( ): lowercase__ = Accelerator() lowercase__ = RegressionDataset(length=80 ) lowercase__ = DataLoader(__magic_name__ , batch_size=16 ) lowercase__ = RegressionDataset(length=96 ) lowercase__ = DataLoader(__magic_name__ , batch_size=16 ) lowercase__ , lowercase__ = accelerator.prepare(__magic_name__ , __magic_name__ ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(__magic_name__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(__magic_name__ ) if iteration < len(__magic_name__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(__magic_name__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(__magic_name__ ) if batch_num < len(__magic_name__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _A ( ): lowercase__ = Accelerator() lowercase__ = accelerator.state if state.local_process_index == 0: print("**Test `accumulate` gradient accumulation with dataloader break**" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("**Test NOOP `no_sync` context manager**" ) test_noop_sync(__magic_name__ ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("**Test Distributed `no_sync` context manager**" ) test_distributed_sync(__magic_name__ ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation, " , f'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation(__magic_name__ , __magic_name__ ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("<" , "2.0" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation with optimizer and scheduler, " , "`split_batches=False`, `dispatch_batches=False`**" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation with optimizer and scheduler, " , f'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation_with_opt_and_scheduler(__magic_name__ , __magic_name__ ) def _A ( __magic_name__ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
655
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase ( lowercase_ ): def __init__( self :Dict , _lowercase :TransformeraDModel , _lowercase :AutoencoderKL , _lowercase :KarrasDiffusionSchedulers , _lowercase :Optional[Dict[int, str]] = None , ): '''simple docstring''' super().__init__() self.register_modules(transformer=_lowercase , vae=_lowercase , scheduler=_lowercase ) # create a imagenet -> id dictionary for easier use lowercase__ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split("," ): lowercase__ = int(_lowercase ) lowercase__ = dict(sorted(self.labels.items() ) ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Union[str, List[str]] ): '''simple docstring''' if not isinstance(_lowercase , _lowercase ): lowercase__ = list(_lowercase ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self :Optional[Any] , _lowercase :List[int] , _lowercase :float = 4.0 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :int = 50 , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = len(_lowercase ) lowercase__ = self.transformer.config.sample_size lowercase__ = self.transformer.config.in_channels lowercase__ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=_lowercase , device=self.device , dtype=self.transformer.dtype , ) lowercase__ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowercase__ = torch.tensor(_lowercase , device=self.device ).reshape(-1 ) lowercase__ = torch.tensor([10_00] * batch_size , device=self.device ) lowercase__ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(_lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowercase__ = latent_model_input[: len(_lowercase ) // 2] lowercase__ = torch.cat([half, half] , dim=0 ) lowercase__ = self.scheduler.scale_model_input(_lowercase , _lowercase ) lowercase__ = t if not torch.is_tensor(_lowercase ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowercase__ = latent_model_input.device.type == "mps" if isinstance(_lowercase , _lowercase ): lowercase__ = torch.floataa if is_mps else torch.floataa else: lowercase__ = torch.intaa if is_mps else torch.intaa lowercase__ = torch.tensor([timesteps] , dtype=_lowercase , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowercase__ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowercase__ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowercase__ = self.transformer( _lowercase , timestep=_lowercase , class_labels=_lowercase ).sample # perform guidance if guidance_scale > 1: lowercase__ , lowercase__ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowercase__ , lowercase__ = torch.split(_lowercase , len(_lowercase ) // 2 , dim=0 ) lowercase__ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowercase__ = torch.cat([half_eps, half_eps] , dim=0 ) lowercase__ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowercase__ , lowercase__ = torch.split(_lowercase , _lowercase , dim=1 ) else: lowercase__ = noise_pred # compute previous image: x_t -> x_t-1 lowercase__ = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample if guidance_scale > 1: lowercase__ , lowercase__ = latent_model_input.chunk(2 , dim=0 ) else: lowercase__ = latent_model_input lowercase__ = 1 / self.vae.config.scaling_factor * latents lowercase__ = self.vae.decode(_lowercase ).sample lowercase__ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowercase__ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (samples,) return ImagePipelineOutput(images=_lowercase )
655
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _snake_case = logging.getLogger(__name__) def _A ( __magic_name__ , __magic_name__ ): return (preds == labels).mean() @dataclass class lowerCAmelCase : __lowerCamelCase = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class lowerCAmelCase : __lowerCamelCase = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) __lowerCamelCase = field(metadata={'help': 'Should contain the data files for the task.'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=lowercase_ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def _A ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( "Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("Training/evaluation parameters %s" , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: lowercase__ = processors[data_args.task_name]() lowercase__ = processor.get_labels() lowercase__ = len(__magic_name__ ) except KeyError: raise ValueError("Task not found: %s" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase__ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) lowercase__ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowercase__ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets lowercase__ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) lowercase__ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: lowercase__ = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator lowercase__ = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowercase__ = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) lowercase__ = trainer.evaluate() lowercase__ = os.path.join(training_args.output_dir , "eval_results.txt" ) if trainer.is_world_master(): with open(__magic_name__ , "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(" %s = %s" , __magic_name__ , __magic_name__ ) writer.write("%s = %s\n" % (key, value) ) results.update(__magic_name__ ) return results def _A ( __magic_name__ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
655
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowerCAmelCase ( lowercase_ ): def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' lowercase__ = SMALL_MODEL_IDENTIFIER lowercase__ = "pt" lowercase__ = "tf" def UpperCAmelCase ( self :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def UpperCAmelCase ( self :Tuple , _lowercase :int ): '''simple docstring''' lowercase__ = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = "mock_framework" # Framework provided - return whatever the user provides lowercase__ = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) lowercase__ = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error lowercase__ = MagicMock(return_value=_lowercase ) lowercase__ = MagicMock(return_value=_lowercase ) with patch("transformers.onnx.features.is_tf_available" , _lowercase ), patch( "transformers.onnx.features.is_torch_available" , _lowercase ): with self.assertRaises(_lowercase ): lowercase__ = FeaturesManager.determine_framework(self.test_model )
655
1
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput _snake_case = """scheduler_config.json""" class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 1 __lowerCamelCase = 2 __lowerCamelCase = 3 __lowerCamelCase = 4 __lowerCamelCase = 5 @dataclass class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 42 class lowerCAmelCase : __lowerCamelCase = SCHEDULER_CONFIG_NAME __lowerCamelCase = ['dtype'] __lowerCamelCase = [] __lowerCamelCase = True @classmethod def UpperCAmelCase ( cls :Optional[Any] , _lowercase :Dict[str, Any] = None , _lowercase :Optional[str] = None , _lowercase :Tuple=False , **_lowercase :Union[str, Any] , ): '''simple docstring''' lowercase__ , lowercase__ = cls.load_config( pretrained_model_name_or_path=_lowercase , subfolder=_lowercase , return_unused_kwargs=_lowercase , **_lowercase , ) lowercase__ , lowercase__ = cls.from_config(_lowercase , return_unused_kwargs=_lowercase , **_lowercase ) if hasattr(_lowercase , "create_state" ) and getattr(_lowercase , "has_state" , _lowercase ): lowercase__ = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCAmelCase ( self :List[str] , _lowercase :Union[str, os.PathLike] , _lowercase :bool = False , **_lowercase :Any ): '''simple docstring''' self.save_config(save_directory=_lowercase , push_to_hub=_lowercase , **_lowercase ) @property def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' return self._get_compatibles() @classmethod def UpperCAmelCase ( cls :Dict ): '''simple docstring''' lowercase__ = list(set([cls.__name__] + cls._compatibles ) ) lowercase__ = importlib.import_module(__name__.split("." )[0] ) lowercase__ = [ getattr(_lowercase , _lowercase ) for c in compatible_classes_str if hasattr(_lowercase , _lowercase ) ] return compatible_classes def _A ( __magic_name__ , __magic_name__ ): assert len(__magic_name__ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__magic_name__ ) - x.ndim) ) , __magic_name__ ) def _A ( __magic_name__ , __magic_name__=0.999 , __magic_name__=jnp.floataa ): def alpha_bar(__magic_name__ ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 lowercase__ = [] for i in range(__magic_name__ ): lowercase__ = i / num_diffusion_timesteps lowercase__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__magic_name__ ) / alpha_bar(__magic_name__ ) , __magic_name__ ) ) return jnp.array(__magic_name__ , dtype=__magic_name__ ) @flax.struct.dataclass class lowerCAmelCase : __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 42 @classmethod def UpperCAmelCase ( cls :Optional[int] , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = scheduler.config if config.trained_betas is not None: lowercase__ = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": lowercase__ = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. lowercase__ = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule lowercase__ = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''' ) lowercase__ = 1.0 - betas lowercase__ = jnp.cumprod(_lowercase , axis=0 ) return cls( alphas=_lowercase , betas=_lowercase , alphas_cumprod=_lowercase , ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state.alphas_cumprod lowercase__ = alphas_cumprod[timesteps] ** 0.5 lowercase__ = sqrt_alpha_prod.flatten() lowercase__ = broadcast_to_shape_from_left(__magic_name__ , original_samples.shape ) lowercase__ = (1 - alphas_cumprod[timesteps]) ** 0.5 lowercase__ = sqrt_one_minus_alpha_prod.flatten() lowercase__ = broadcast_to_shape_from_left(__magic_name__ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = get_sqrt_alpha_prod(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = get_sqrt_alpha_prod(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
655
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git_vision_model' def __init__( self :Dict , _lowercase :Dict=7_68 , _lowercase :Dict=30_72 , _lowercase :Tuple=12 , _lowercase :List[str]=12 , _lowercase :Tuple=3 , _lowercase :Dict=2_24 , _lowercase :Tuple=16 , _lowercase :Optional[int]="quick_gelu" , _lowercase :Union[str, Any]=1e-5 , _lowercase :Tuple=0.0 , _lowercase :Tuple=0.02 , **_lowercase :Optional[Any] , ): '''simple docstring''' super().__init__(**_lowercase ) lowercase__ = hidden_size lowercase__ = intermediate_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = num_channels lowercase__ = patch_size lowercase__ = image_size lowercase__ = initializer_range lowercase__ = attention_dropout lowercase__ = layer_norm_eps lowercase__ = hidden_act @classmethod def UpperCAmelCase ( cls :List[str] , _lowercase :Union[str, os.PathLike] , **_lowercase :Optional[int] ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class lowerCAmelCase ( lowercase_ ): __lowerCamelCase = 'git' def __init__( self :Union[str, Any] , _lowercase :Dict=None , _lowercase :List[str]=3_05_22 , _lowercase :Tuple=7_68 , _lowercase :Any=6 , _lowercase :Dict=12 , _lowercase :Any=30_72 , _lowercase :List[Any]="gelu" , _lowercase :Tuple=0.1 , _lowercase :Optional[int]=0.1 , _lowercase :Optional[Any]=10_24 , _lowercase :Any=0.02 , _lowercase :int=1e-12 , _lowercase :List[Any]=0 , _lowercase :int="absolute" , _lowercase :List[str]=True , _lowercase :Any=False , _lowercase :int=1_01 , _lowercase :str=1_02 , _lowercase :Dict=None , **_lowercase :List[str] , ): '''simple docstring''' super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , pad_token_id=_lowercase , **_lowercase ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowercase__ = GitVisionConfig(**_lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = tie_word_embeddings lowercase__ = num_image_with_embedding lowercase__ = bos_token_id lowercase__ = eos_token_id def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
655
1
from collections import deque class lowerCAmelCase : def __init__( self :List[Any] , _lowercase :str , _lowercase :int , _lowercase :int ): '''simple docstring''' lowercase__ = process_name # process name lowercase__ = arrival_time # arrival time of the process # completion time of finished process or last interrupted time lowercase__ = arrival_time lowercase__ = burst_time # remaining burst time lowercase__ = 0 # total time of the process wait in ready queue lowercase__ = 0 # time from arrival time to completion time class lowerCAmelCase : def __init__( self :Optional[int] , _lowercase :int , _lowercase :list[int] , _lowercase :deque[Process] , _lowercase :int , ): '''simple docstring''' lowercase__ = number_of_queues # time slice of queues that round robin algorithm applied lowercase__ = time_slices # unfinished process is in this ready_queue lowercase__ = queue # current time lowercase__ = current_time # finished process is in this sequence queue lowercase__ = deque() def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def UpperCAmelCase ( self :List[str] , _lowercase :list[Process] ): '''simple docstring''' lowercase__ = [] for i in range(len(_lowercase ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def UpperCAmelCase ( self :Dict , _lowercase :list[Process] ): '''simple docstring''' lowercase__ = [] for i in range(len(_lowercase ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def UpperCAmelCase ( self :Dict , _lowercase :list[Process] ): '''simple docstring''' lowercase__ = [] for i in range(len(_lowercase ) ): completion_times.append(queue[i].stop_time ) return completion_times def UpperCAmelCase ( self :Optional[Any] , _lowercase :deque[Process] ): '''simple docstring''' return [q.burst_time for q in queue] def UpperCAmelCase ( self :Optional[int] , _lowercase :Process ): '''simple docstring''' process.waiting_time += self.current_time - process.stop_time return process.waiting_time def UpperCAmelCase ( self :List[str] , _lowercase :deque[Process] ): '''simple docstring''' lowercase__ = deque() # sequence deque of finished process while len(_lowercase ) != 0: lowercase__ = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(_lowercase ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 lowercase__ = 0 # set the process's turnaround time because it is finished lowercase__ = self.current_time - cp.arrival_time # set the completion time lowercase__ = self.current_time # add the process to queue that has finished queue finished.append(_lowercase ) self.finish_queue.extend(_lowercase ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def UpperCAmelCase ( self :Any , _lowercase :deque[Process] , _lowercase :int ): '''simple docstring''' lowercase__ = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(_lowercase ) ): lowercase__ = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(_lowercase ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time lowercase__ = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(_lowercase ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished lowercase__ = 0 # set the finish time lowercase__ = self.current_time # update the process' turnaround time because it is finished lowercase__ = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(_lowercase ) self.finish_queue.extend(_lowercase ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for i in range(self.number_of_queues - 1 ): lowercase__ , lowercase__ = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest _snake_case = Process("""P1""", 0, 53) _snake_case = Process("""P2""", 0, 17) _snake_case = Process("""P3""", 0, 68) _snake_case = Process("""P4""", 0, 24) _snake_case = 3 _snake_case = [17, 25] _snake_case = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={"""queue""": deque([Pa, Pa, Pa, Pa])}) _snake_case = Process("""P1""", 0, 53) _snake_case = Process("""P2""", 0, 17) _snake_case = Process("""P3""", 0, 68) _snake_case = Process("""P4""", 0, 24) _snake_case = 3 _snake_case = [17, 25] _snake_case = deque([Pa, Pa, Pa, Pa]) _snake_case = MLFQ(number_of_queues, time_slices, queue, 0) _snake_case = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F"""waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print completion times of processes(P1, P2, P3, P4) print( F"""completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print total turnaround times of processes(P1, P2, P3, P4) print( F"""turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print sequence of finished processes print( F"""sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}""" )
655
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class lowerCAmelCase ( unittest.TestCase ): @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModel.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModel.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForPreTraining.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForCausalLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForCausalLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :List[str] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForMaskedLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForMaskedLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :Tuple ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_pt=_lowercase ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSeqaSeqLM.from_pretrained(_lowercase , from_tf=_lowercase ) lowercase__ , lowercase__ = AutoModelForSeqaSeqLM.from_pretrained( _lowercase , output_loading_info=_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForSequenceClassification.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) @slow def UpperCAmelCase ( self :str ): '''simple docstring''' for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) lowercase__ = AutoModelForQuestionAnswering.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsNotNone(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = TFAutoModelWithLMHead.from_pretrained(_lowercase , from_pt=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 ) lowercase__ = AutoModelWithLMHead.from_pretrained(_lowercase , from_tf=_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=_lowercase ) , 1_44_10 )
655
1
from __future__ import annotations _snake_case = 8.988E9 # units = N * m^s * C^-2 def _A ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if distance < 0: raise ValueError("Distance cannot be negative" ) if force == 0: lowercase__ = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: lowercase__ = abs(__magic_name__ ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: lowercase__ = abs(__magic_name__ ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: lowercase__ = (COULOMBS_CONSTANT * charge_product / abs(__magic_name__ )) ** 0.5 return {"distance": distance} raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
655
import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _snake_case = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val def _A ( __magic_name__ ): lowercase__ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowercase__ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) lowercase__ = value else: lowercase__ = value return new_state_dict def _A ( __magic_name__ ): lowercase__ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase__ = in_proj_weight[:256, :] lowercase__ = in_proj_bias[:256] lowercase__ = in_proj_weight[256:512, :] lowercase__ = in_proj_bias[256:512] lowercase__ = in_proj_weight[-256:, :] lowercase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowercase__ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) lowercase__ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowercase__ = in_proj_weight_cross_attn[:256, :] lowercase__ = in_proj_bias_cross_attn[:256] lowercase__ = in_proj_weight_cross_attn[256:512, :] lowercase__ = in_proj_bias_cross_attn[256:512] lowercase__ = in_proj_weight_cross_attn[-256:, :] lowercase__ = in_proj_bias_cross_attn[-256:] def _A ( __magic_name__ , __magic_name__ ): lowercase__ , lowercase__ = image.size lowercase__ = max(__magic_name__ , __magic_name__ ) lowercase__ = 800 if "detection" in checkpoint_url else 1000 lowercase__ = target_max_size / current_max_size lowercase__ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _A ( __magic_name__ ): lowercase__ = F.to_tensor(__magic_name__ ) lowercase__ = F.normalize(__magic_name__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def _A ( __magic_name__ , __magic_name__ , __magic_name__ ): logger.info("Converting model..." ) # load original state dict lowercase__ = torch.hub.load_state_dict_from_url(__magic_name__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ , __magic_name__ , __magic_name__ ) lowercase__ = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowercase__ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowercase__ = state_dict.pop(__magic_name__ ) lowercase__ = val # create HuggingFace model and load state dict lowercase__ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: lowercase__ = 15 lowercase__ = 2 lowercase__ = {0: "table", 1: "table rotated"} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} else: lowercase__ = 125 lowercase__ = 6 lowercase__ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) lowercase__ = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion lowercase__ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" lowercase__ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=__magic_name__ ) lowercase__ = Image.open(__magic_name__ ).convert("RGB" ) lowercase__ = normalize(resize(__magic_name__ , __magic_name__ ) ).unsqueeze(0 ) lowercase__ = model(__magic_name__ ) if "detection" in checkpoint_url: lowercase__ = (1, 15, 3) lowercase__ = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) lowercase__ = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: lowercase__ = (1, 125, 7) lowercase__ = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) lowercase__ = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __magic_name__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) lowercase__ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _snake_case = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
655
1
from __future__ import annotations import time _snake_case = list[tuple[int, int]] _snake_case = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _snake_case = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class lowerCAmelCase : def __init__( self :List[Any] , _lowercase :int , _lowercase :int , _lowercase :int , _lowercase :int , _lowercase :Node | None ): '''simple docstring''' lowercase__ = pos_x lowercase__ = pos_y lowercase__ = (pos_y, pos_x) lowercase__ = goal_x lowercase__ = goal_y lowercase__ = parent class lowerCAmelCase : def __init__( self :Any , _lowercase :tuple[int, int] , _lowercase :tuple[int, int] ): '''simple docstring''' lowercase__ = Node(start[1] , start[0] , goal[1] , goal[0] , _lowercase ) lowercase__ = Node(goal[1] , goal[0] , goal[1] , goal[0] , _lowercase ) lowercase__ = [self.start] lowercase__ = False def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' while self.node_queue: lowercase__ = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: lowercase__ = True return self.retrace_path(_lowercase ) lowercase__ = self.get_successors(_lowercase ) for node in successors: self.node_queue.append(_lowercase ) if not self.reached: return [self.start.pos] return None def UpperCAmelCase ( self :Tuple , _lowercase :Node ): '''simple docstring''' lowercase__ = [] for action in delta: lowercase__ = parent.pos_x + action[1] lowercase__ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(_lowercase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(_lowercase , _lowercase , self.target.pos_y , self.target.pos_x , _lowercase ) ) return successors def UpperCAmelCase ( self :List[Any] , _lowercase :Node | None ): '''simple docstring''' lowercase__ = node lowercase__ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) lowercase__ = current_node.parent path.reverse() return path class lowerCAmelCase : def __init__( self :List[str] , _lowercase :str , _lowercase :List[Any] ): '''simple docstring''' lowercase__ = BreadthFirstSearch(_lowercase , _lowercase ) lowercase__ = BreadthFirstSearch(_lowercase , _lowercase ) lowercase__ = False def UpperCAmelCase ( self :Dict ): '''simple docstring''' while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: lowercase__ = self.fwd_bfs.node_queue.pop(0 ) lowercase__ = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: lowercase__ = True return self.retrace_bidirectional_path( _lowercase , _lowercase ) lowercase__ = current_bwd_node lowercase__ = current_fwd_node lowercase__ = { self.fwd_bfs: self.fwd_bfs.get_successors(_lowercase ), self.bwd_bfs: self.bwd_bfs.get_successors(_lowercase ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(_lowercase ) if not self.reached: return [self.fwd_bfs.start.pos] return None def UpperCAmelCase ( self :Optional[int] , _lowercase :Node , _lowercase :Node ): '''simple docstring''' lowercase__ = self.fwd_bfs.retrace_path(_lowercase ) lowercase__ = self.bwd_bfs.retrace_path(_lowercase ) bwd_path.pop() bwd_path.reverse() lowercase__ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() _snake_case = (0, 0) _snake_case = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _snake_case = time.time() _snake_case = BreadthFirstSearch(init, goal) _snake_case = bfs.search() _snake_case = time.time() - start_bfs_time print("""Unidirectional BFS computation time : """, bfs_time) _snake_case = time.time() _snake_case = BidirectionalBreadthFirstSearch(init, goal) _snake_case = bd_bfs.search() _snake_case = time.time() - start_bd_bfs_time print("""Bidirectional BFS computation time : """, bd_bfs_time)
655
from typing import TYPE_CHECKING from ...utils import _LazyModule _snake_case = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
655
1
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase ( lowercase_ , unittest.TestCase ): __lowerCamelCase = ConsistencyModelPipeline __lowerCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS __lowerCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt __lowerCamelCase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet" , ) return unet @property def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet_class_cond" , ) return unet def UpperCAmelCase ( self :str , _lowercase :Any=False ): '''simple docstring''' if class_cond: lowercase__ = self.dummy_cond_unet else: lowercase__ = self.dummy_uncond_unet # Default to CM multistep sampler lowercase__ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) lowercase__ = { "unet": unet, "scheduler": scheduler, } return components def UpperCAmelCase ( self :Any , _lowercase :Optional[int] , _lowercase :Any=0 ): '''simple docstring''' if str(_lowercase ).startswith("mps" ): lowercase__ = torch.manual_seed(_lowercase ) else: lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = { "batch_size": 1, "num_inference_steps": None, "timesteps": [22, 0], "generator": generator, "output_type": "np", } return inputs def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = ConsistencyModelPipeline(**_lowercase ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 32, 32, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components(class_cond=_lowercase ) lowercase__ = ConsistencyModelPipeline(**_lowercase ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = 0 lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 32, 32, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = ConsistencyModelPipeline(**_lowercase ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = 1 lowercase__ = None lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 32, 32, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components(class_cond=_lowercase ) lowercase__ = ConsistencyModelPipeline(**_lowercase ) lowercase__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = 1 lowercase__ = None lowercase__ = 0 lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 32, 32, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): def UpperCAmelCase ( self :List[str] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self :str , _lowercase :Any=0 , _lowercase :Tuple=False , _lowercase :List[str]="cpu" , _lowercase :List[str]=torch.floataa , _lowercase :Optional[Any]=(1, 3, 64, 64) ): '''simple docstring''' lowercase__ = torch.manual_seed(_lowercase ) lowercase__ = { "num_inference_steps": None, "timesteps": [22, 0], "class_labels": 0, "generator": generator, "output_type": "np", } if get_fixed_latents: lowercase__ = self.get_fixed_latents(seed=_lowercase , device=_lowercase , dtype=_lowercase , shape=_lowercase ) lowercase__ = latents return inputs def UpperCAmelCase ( self :Dict , _lowercase :Tuple=0 , _lowercase :List[Any]="cpu" , _lowercase :Optional[int]=torch.floataa , _lowercase :List[str]=(1, 3, 64, 64) ): '''simple docstring''' if type(_lowercase ) == str: lowercase__ = torch.device(_lowercase ) lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) return latents def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) lowercase__ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) lowercase__ = ConsistencyModelPipeline(unet=_lowercase , scheduler=_lowercase ) pipe.to(torch_device=_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_inputs() lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 64, 64, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.0888, 0.0881, 0.0666, 0.0479, 0.0292, 0.0195, 0.0201, 0.0163, 0.0254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) lowercase__ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) lowercase__ = ConsistencyModelPipeline(unet=_lowercase , scheduler=_lowercase ) pipe.to(torch_device=_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_inputs() lowercase__ = 1 lowercase__ = None lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 64, 64, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.0340, 0.0152, 0.0063, 0.0267, 0.0221, 0.0107, 0.0416, 0.0186, 0.0217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) lowercase__ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) lowercase__ = ConsistencyModelPipeline(unet=_lowercase , scheduler=_lowercase ) pipe.to(torch_device=_lowercase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_inputs(get_fixed_latents=_lowercase , device=_lowercase ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=_lowercase , enable_math=_lowercase , enable_mem_efficient=_lowercase ): lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 64, 64, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.1875, 0.1428, 0.1289, 0.2151, 0.2092, 0.1477, 0.1877, 0.1641, 0.1353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) lowercase__ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) lowercase__ = ConsistencyModelPipeline(unet=_lowercase , scheduler=_lowercase ) pipe.to(torch_device=_lowercase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_inputs(get_fixed_latents=_lowercase , device=_lowercase ) lowercase__ = 1 lowercase__ = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=_lowercase , enable_math=_lowercase , enable_mem_efficient=_lowercase ): lowercase__ = pipe(**_lowercase ).images assert image.shape == (1, 64, 64, 3) lowercase__ = image[0, -3:, -3:, -1] lowercase__ = np.array([0.1663, 0.1948, 0.2275, 0.1680, 0.1204, 0.1245, 0.1858, 0.1338, 0.2095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
655
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case = """ Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\") >>> pipe.to(\"cuda\") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _A ( __magic_name__ , __magic_name__ , __magic_name__=8 ): lowercase__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCAmelCase ( lowercase_ ): def __init__( self :List[str] , _lowercase :UNetaDConditionModel , _lowercase :DDPMScheduler , _lowercase :VQModel , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) lowercase__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self :Union[str, Any] , _lowercase :Tuple , _lowercase :List[str] , _lowercase :Tuple , _lowercase :Optional[Any] , _lowercase :int , _lowercase :str ): '''simple docstring''' if latents is None: lowercase__ = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase__ = latents.to(_lowercase ) lowercase__ = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self :int , _lowercase :int=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) lowercase__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def UpperCAmelCase ( self :Optional[int] , _lowercase :Tuple=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) lowercase__ = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase__ = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase__ , lowercase__ = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. lowercase__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__( self :int , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :Union[torch.FloatTensor, List[torch.FloatTensor]] , _lowercase :int = 5_12 , _lowercase :int = 5_12 , _lowercase :int = 1_00 , _lowercase :float = 4.0 , _lowercase :int = 1 , _lowercase :Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowercase :Optional[torch.FloatTensor] = None , _lowercase :Optional[str] = "pil" , _lowercase :bool = True , ): '''simple docstring''' lowercase__ = self._execution_device lowercase__ = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) lowercase__ = image_embeds.shape[0] * num_images_per_prompt if isinstance(_lowercase , _lowercase ): lowercase__ = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: lowercase__ = image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) lowercase__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) lowercase__ = self.scheduler.timesteps lowercase__ = self.unet.config.in_channels lowercase__ , lowercase__ = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) # create initial latent lowercase__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _lowercase , _lowercase , _lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance lowercase__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase__ = {"image_embeds": image_embeds} lowercase__ = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) lowercase__ , lowercase__ = noise_pred.chunk(2 ) lowercase__ , lowercase__ = variance_pred.chunk(2 ) lowercase__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase__ , lowercase__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase__ = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing lowercase__ = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase__ = image * 0.5 + 0.5 lowercase__ = image.clamp(0 , 1 ) lowercase__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase__ = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
655
1