code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> str: super().__init__( features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ , streaming=lowerCAmelCase__ , num_proc=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : str = Generator( cache_dir=lowerCAmelCase__ , features=lowerCAmelCase__ , generator=lowerCAmelCase__ , gen_kwargs=lowerCAmelCase__ , **lowerCAmelCase__ , ) def __a ( self ) -> Union[str, Any]: # Build iterable dataset if self.streaming: a : Optional[int] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: a : List[Any] = None a : Tuple = None a : Union[str, Any] = None a : int = None self.builder.download_and_prepare( download_config=lowerCAmelCase__ , download_mode=lowerCAmelCase__ , verification_mode=lowerCAmelCase__ , base_path=lowerCAmelCase__ , num_proc=self.num_proc , ) a : List[str] = self.builder.as_dataset( split="train" , verification_mode=lowerCAmelCase__ , in_memory=self.keep_in_memory ) return dataset
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py a : Optional[int] = '''src/transformers''' a : List[str] = '''docs/source/en/tasks''' def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : int , _lowercase : int ) ->List[Any]: '''simple docstring''' with open(_lowercase , "r" , encoding="utf-8" , newline="\n" ) as f: a : int = f.readlines() # Find the start prompt. a : Dict = 0 while not lines[start_index].startswith(_lowercase ): start_index += 1 start_index += 1 a : int = start_index while not lines[end_index].startswith(_lowercase ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. a : Any = direct_transformers_import(TRANSFORMERS_PATH) a : List[Any] = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). a : Optional[Any] = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->Optional[Any]: '''simple docstring''' a : str = TASK_GUIDE_TO_MODELS[task_guide] a : int = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(_lowercase , set() ) a : Optional[Any] = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"""[{name}](../model_doc/{code})""" for code, name in model_names.items()] ) + "\n" def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : int=False ) ->Union[str, Any]: '''simple docstring''' a, a, a, a : str = _find_text_in_file( filename=os.path.join(_lowercase , _lowercase ) , start_prompt="<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->" , end_prompt="<!--End of the generated tip-->" , ) a : Union[str, Any] = get_model_list_for_task(_lowercase ) if current_list != new_list: if overwrite: with open(os.path.join(_lowercase , _lowercase ) , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"""The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`""" " to fix this." ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') a : Dict = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = {'''configuration_mmbt''': ['''MMBTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = ['''MMBTForClassification''', '''MMBTModel''', '''ModalEmbeddings'''] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys a : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version a : Dict = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.8.0''', '''To fix: pip install -r examples/pytorch/image-classification/requirements.txt''') a : Any = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) a : Tuple = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' with open(_lowercase , "rb" ) as f: a : List[str] = Image.open(_lowercase ) return im.convert("RGB" ) @dataclass class __UpperCamelCase : lowerCamelCase : Optional[str] =field( default=a__ , metadata={ """help""": """Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub).""" } , ) lowerCamelCase : Optional[str] =field( default=a__ , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) lowerCamelCase : Optional[str] =field(default=a__ , metadata={"""help""": """A folder containing the training data."""} ) lowerCamelCase : Optional[str] =field(default=a__ , metadata={"""help""": """A folder containing the validation data."""} ) lowerCamelCase : Optional[float] =field( default=0.15 , metadata={"""help""": """Percent to split off of train for validation."""} ) lowerCamelCase : Optional[int] =field( default=a__ , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowerCamelCase : Optional[int] =field( default=a__ , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __a ( self ) -> List[Any]: if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class __UpperCamelCase : lowerCamelCase : str =field( default="""google/vit-base-patch16-224-in21k""" , metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} , ) lowerCamelCase : Optional[str] =field( default=a__ , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(a__ )} , ) lowerCamelCase : Optional[str] =field( default=a__ , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowerCamelCase : Optional[str] =field( default=a__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from s3"""} ) lowerCamelCase : str =field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowerCamelCase : str =field(default=a__ , metadata={"""help""": """Name or path of preprocessor config."""} ) lowerCamelCase : bool =field( default=a__ , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) lowerCamelCase : bool =field( default=a__ , metadata={"""help""": """Will enable to load a pretrained model whose head dimensions are different."""} , ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->Dict: '''simple docstring''' a : Tuple = torch.stack([example["pixel_values"] for example in examples] ) a : Dict = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def _SCREAMING_SNAKE_CASE ( ) ->Any: '''simple docstring''' a : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. a, a, a : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: a, a, a : int = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , _lowercase , _lowercase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() a : Optional[int] = training_args.get_process_log_level() logger.setLevel(_lowercase ) transformers.utils.logging.set_verbosity(_lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. a : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: a : Optional[int] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: a : Tuple = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: a : Tuple = {} if data_args.train_dir is not None: a : int = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: a : Any = os.path.join(data_args.validation_dir , "**" ) a : Optional[Any] = load_dataset( "imagefolder" , data_files=_lowercase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. a : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _lowercase ) and data_args.train_val_split > 0.0: a : int = dataset["train"].train_test_split(data_args.train_val_split ) a : str = split["train"] a : Dict = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. a : List[str] = dataset["train"].features["labels"].names a, a : Union[str, Any] = {}, {} for i, label in enumerate(_lowercase ): a : int = str(_lowercase ) a : List[Any] = label # Load the accuracy metric from the datasets package a : Union[str, Any] = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(_lowercase : Optional[int] ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) a : Any = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(_lowercase ) , labelaid=_lowercase , idalabel=_lowercase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) a : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=_lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) a : List[Any] = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: a : str = image_processor.size["shortest_edge"] else: a : List[str] = (image_processor.size["height"], image_processor.size["width"]) a : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) a : int = Compose( [ RandomResizedCrop(_lowercase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) a : List[str] = Compose( [ Resize(_lowercase ), CenterCrop(_lowercase ), ToTensor(), normalize, ] ) def train_transforms(_lowercase : List[str] ): a : List[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(_lowercase : List[Any] ): a : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: a : Optional[int] = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(_lowercase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: a : Any = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(_lowercase ) # Initalize our trainer a : Union[str, Any] = Trainer( model=_lowercase , args=_lowercase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=_lowercase , tokenizer=_lowercase , data_collator=_lowercase , ) # Training if training_args.do_train: a : List[str] = None if training_args.resume_from_checkpoint is not None: a : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: a : Dict = last_checkpoint a : List[str] = trainer.train(resume_from_checkpoint=_lowercase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: a : Dict = trainer.evaluate() trainer.log_metrics("eval" , _lowercase ) trainer.save_metrics("eval" , _lowercase ) # Write model card and (optionally) push to hub a : List[str] = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**_lowercase ) else: trainer.create_model_card(**_lowercase ) if __name__ == "__main__": main()
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) a : int = logging.get_logger(__name__) # pylint: disable=invalid-name a : Optional[int] = ''' Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior") >>> pipe_prior.to("cuda") >>> prompt = "red cat, 4k photo" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder") >>> pipe.to("cuda") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save("cat.png") ``` ''' def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : List[Any] , _lowercase : Optional[int]=8 ) ->List[str]: '''simple docstring''' a : str = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 a : int = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> Tuple: super().__init__() self.register_modules( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , movq=lowerCAmelCase__ , ) a : str = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: if latents is None: a : List[Any] = randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) a : str = latents.to(lowerCAmelCase__ ) a : List[Any] = latents * scheduler.init_noise_sigma return latents def __a ( self , lowerCAmelCase__=0 ) -> List[str]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) a : Optional[int] = torch.device(f"""cuda:{gpu_id}""" ) a : str = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__=0 ) -> Tuple: if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) a : Union[str, Any] = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowerCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) a : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: a, a : Any = cpu_offload_with_hook(lowerCAmelCase__ , lowerCAmelCase__ , prev_module_hook=lowerCAmelCase__ ) # We'll offload the last model manually. a : Union[str, Any] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __a ( self ) -> Optional[int]: if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCAmelCase__ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowerCAmelCase__ ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 512 , lowerCAmelCase__ = 512 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 4.0 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = "pil" , lowerCAmelCase__ = True , ) -> List[str]: a : Any = self._execution_device a : List[str] = guidance_scale > 1.0 if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : str = torch.cat(lowerCAmelCase__ , dim=0 ) a : List[Any] = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[str] = torch.cat(lowerCAmelCase__ , dim=0 ) if do_classifier_free_guidance: a : Optional[int] = image_embeds.repeat_interleave(lowerCAmelCase__ , dim=0 ) a : Any = negative_image_embeds.repeat_interleave(lowerCAmelCase__ , dim=0 ) a : int = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowerCAmelCase__ ) self.scheduler.set_timesteps(lowerCAmelCase__ , device=lowerCAmelCase__ ) a : Any = self.scheduler.timesteps a : Any = self.unet.config.in_channels a, a : Dict = downscale_height_and_width(lowerCAmelCase__ , lowerCAmelCase__ , self.movq_scale_factor ) # create initial latent a : Union[str, Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowerCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance a : Optional[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a : Any = {"image_embeds": image_embeds} a : List[Any] = self.unet( sample=lowerCAmelCase__ , timestep=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , added_cond_kwargs=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] if do_classifier_free_guidance: a, a : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) a, a : str = noise_pred.chunk(2 ) a, a : str = variance_pred.chunk(2 ) a : List[str] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) a : Dict = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): a, a : Any = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 a : Tuple = self.scheduler.step( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , generator=lowerCAmelCase__ , )[0] # post-processing a : Union[str, Any] = self.movq.decode(lowerCAmelCase__ , force_not_quantize=lowerCAmelCase__ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: a : Tuple = image * 0.5 + 0.5 a : int = image.clamp(0 , 1 ) a : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": a : Optional[int] = self.numpy_to_pil(lowerCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCAmelCase__ )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class __UpperCamelCase ( a__ ): lowerCamelCase : int ="""M-CLIP""" def __init__( self , lowerCAmelCase__=1024 , lowerCAmelCase__=768 , **lowerCAmelCase__ ) -> Optional[Any]: a : str = transformerDimSize a : Any = imageDimSize super().__init__(**lowerCAmelCase__ ) class __UpperCamelCase ( a__ ): lowerCamelCase : str =MCLIPConfig def __init__( self , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: super().__init__(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) a : List[str] = XLMRobertaModel(lowerCAmelCase__ ) a : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: a : int = self.transformer(input_ids=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] a : str = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(lowerCAmelCase__ ), embs
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : str = { '''configuration_autoformer''': [ '''AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AutoformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ '''AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AutoformerForPrediction''', '''AutoformerModel''', '''AutoformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Union[str, Any] =BertJapaneseTokenizer lowerCamelCase : List[Any] =False lowerCamelCase : List[Any] =True def __a ( self ) -> List[Any]: super().setUp() a : Optional[int] = [ "[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは", "世界", "##世界", "、", "##、", "。", "##。", ] a : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: a : Union[str, Any] = "こんにちは、世界。 \nこんばんは、世界。" a : Any = "こんにちは 、 世界 。 こんばんは 、 世界 。" return input_text, output_text def __a ( self , lowerCAmelCase__ ) -> Any: a, a : Tuple = self.get_input_output_texts(lowerCAmelCase__ ) a : List[str] = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) a : Optional[Any] = tokenizer.decode(lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ ) return text, ids def __a ( self ) -> str: pass # TODO add if relevant def __a ( self ) -> str: pass # TODO add if relevant def __a ( self ) -> str: pass # TODO add if relevant def __a ( self ) -> Optional[Any]: a : List[Any] = self.tokenizer_class(self.vocab_file ) a : Optional[int] = tokenizer.tokenize("こんにちは、世界。\nこんばんは、世界。" ) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __a ( self ) -> Optional[Any]: a : List[str] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="mecab" ) self.assertIsNotNone(lowerCAmelCase__ ) a : int = "こんにちは、世界。\nこんばんは、世界。" a : List[str] = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) a : int = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(lowerCAmelCase__ , "wb" ) as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__ ) with open(lowerCAmelCase__ , "rb" ) as handle: a : List[Any] = pickle.load(lowerCAmelCase__ ) a : Union[str, Any] = tokenizer_new.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : Any = MecabTokenizer(mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __a ( self ) -> str: try: a : List[str] = MecabTokenizer(mecab_dic="unidic_lite" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __a ( self ) -> Tuple: try: a : Union[str, Any] = MecabTokenizer(mecab_dic="unidic" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __a ( self ) -> List[Any]: a : int = MecabTokenizer(do_lower_case=lowerCAmelCase__ , mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iphone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) def __a ( self ) -> Union[str, Any]: try: a : List[str] = MecabTokenizer( do_lower_case=lowerCAmelCase__ , normalize_text=lowerCAmelCase__ , mecab_option="-d /usr/local/lib/mecab/dic/jumandic" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "\u3000", "。"] , ) def __a ( self ) -> Optional[Any]: a : int = MecabTokenizer(normalize_text=lowerCAmelCase__ , mecab_dic="ipadic" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップルストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", " ", "。"] , ) @require_sudachi def __a ( self ) -> Union[str, Any]: a : Union[str, Any] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="sudachi" ) self.assertIsNotNone(lowerCAmelCase__ ) a : int = "こんにちは、世界。\nこんばんは、世界。" a : Dict = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) a : Any = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(lowerCAmelCase__ , "wb" ) as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__ ) with open(lowerCAmelCase__ , "rb" ) as handle: a : Any = pickle.load(lowerCAmelCase__ ) a : Optional[Any] = tokenizer_new.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) @require_sudachi def __a ( self ) -> Any: a : Tuple = SudachiTokenizer(sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def __a ( self ) -> Dict: a : int = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="A" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国", "人", "参政", "権"] ) @require_sudachi def __a ( self ) -> int: a : int = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="B" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国人", "参政権"] ) @require_sudachi def __a ( self ) -> Dict: a : Optional[Any] = SudachiTokenizer(sudachi_dict_type="core" , sudachi_split_mode="C" ) self.assertListEqual(tokenizer.tokenize("外国人参政権" ) , ["外国人参政権"] ) @require_sudachi def __a ( self ) -> Tuple: a : Union[str, Any] = SudachiTokenizer(do_lower_case=lowerCAmelCase__ , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iphone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", " ", "。", " ", " "] , ) @require_sudachi def __a ( self ) -> str: a : List[Any] = SudachiTokenizer(normalize_text=lowerCAmelCase__ , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , [" ", "\t", "アップル", "ストア", "で", "iPhone", "8", " ", "が", " ", " ", "\n ", "発売", "さ", "れ", "た", "\u3000", "。", " ", " "] , ) @require_sudachi def __a ( self ) -> List[str]: a : List[Any] = SudachiTokenizer(trim_whitespace=lowerCAmelCase__ , sudachi_dict_type="core" ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れ", "た", "。"] , ) @require_jumanpp def __a ( self ) -> Tuple: a : List[str] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="jumanpp" ) self.assertIsNotNone(lowerCAmelCase__ ) a : str = "こんにちは、世界。\nこんばんは、世界。" a : List[str] = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , ["こんにちは", "、", "世界", "。", "こん", "##ばんは", "、", "世界", "。"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) a : List[Any] = os.path.join(self.tmpdirname , "tokenizer.bin" ) with open(lowerCAmelCase__ , "wb" ) as handle: pickle.dump(lowerCAmelCase__ , lowerCAmelCase__ ) with open(lowerCAmelCase__ , "rb" ) as handle: a : str = pickle.load(lowerCAmelCase__ ) a : List[str] = tokenizer_new.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) @require_jumanpp def __a ( self ) -> List[str]: a : int = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __a ( self ) -> Union[str, Any]: a : Optional[Any] = JumanppTokenizer(do_lower_case=lowerCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iphone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __a ( self ) -> Any: a : List[Any] = JumanppTokenizer(normalize_text=lowerCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["ア", "ッ", "フ", "゚", "ル", "ストア", "で", "iPhone", "8", "\u3000", "が", "\u3000", "\u3000", "\u3000", "発売", "さ", "れた", "\u3000", "。"] , ) @require_jumanpp def __a ( self ) -> Any: a : Tuple = JumanppTokenizer(trim_whitespace=lowerCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tアップルストアでiPhone8 が \n 発売された 。 " ) , ["アップル", "ストア", "で", "iPhone", "8", "が", "発売", "さ", "れた", "。"] , ) @require_jumanpp def __a ( self ) -> Tuple: a : Optional[Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("ありがとうございますm(_ _)m見つけるのが大変です。" ) , ["ありがとう", "ございます", "m(_ _)m", "見つける", "の", "が", "大変です", "。"] , ) def __a ( self ) -> str: a : Optional[Any] = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは"] a : Optional[Any] = {} for i, token in enumerate(lowerCAmelCase__ ): a : int = i a : Optional[Any] = WordpieceTokenizer(vocab=lowerCAmelCase__ , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) , ["こんにちは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは" ) , ["こん", "##ばんは"] ) self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは" ) , ["こん", "##ばんは", "[UNK]", "こんにちは"] ) def __a ( self ) -> int: a : List[str] = BertJapaneseTokenizer.from_pretrained("nlp-waseda/roberta-base-japanese-with-auto-jumanpp" ) a : Tuple = tokenizer.subword_tokenizer a : List[str] = subword_tokenizer.tokenize("国境 の 長い トンネル を 抜ける と 雪国 であった 。" ) self.assertListEqual(lowerCAmelCase__ , ["▁国境", "▁の", "▁長い", "▁トンネル", "▁を", "▁抜ける", "▁と", "▁雪", "国", "▁であった", "▁。"] ) a : Dict = subword_tokenizer.tokenize("こんばんは こんばん にち は こんにちは" ) self.assertListEqual(lowerCAmelCase__ , ["▁こん", "ばん", "は", "▁こん", "ばん", "▁に", "ち", "▁は", "▁こんにちは"] ) def __a ( self ) -> Any: a : int = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese" ) a : List[Any] = tokenizer.encode("ありがとう。" , add_special_tokens=lowerCAmelCase__ ) a : Union[str, Any] = tokenizer.encode("どういたしまして。" , add_special_tokens=lowerCAmelCase__ ) a : List[Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ) a : Any = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Any =BertJapaneseTokenizer lowerCamelCase : Tuple =False def __a ( self ) -> Tuple: super().setUp() a : Union[str, Any] = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] a : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __a ( self , **lowerCAmelCase__ ) -> str: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="character" , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: a : List[str] = "こんにちは、世界。 \nこんばんは、世界。" a : int = "こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。" return input_text, output_text def __a ( self ) -> int: pass # TODO add if relevant def __a ( self ) -> Optional[Any]: pass # TODO add if relevant def __a ( self ) -> Any: pass # TODO add if relevant def __a ( self ) -> Dict: a : int = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="character" ) a : List[Any] = tokenizer.tokenize("こんにちは、世界。 \nこんばんは、世界。" ) self.assertListEqual( lowerCAmelCase__ , ["こ", "ん", "に", "ち", "は", "、", "世", "界", "。", "こ", "ん", "ば", "ん", "は", "、", "世", "界", "。"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __a ( self ) -> str: a : Union[str, Any] = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"] a : Optional[int] = {} for i, token in enumerate(lowerCAmelCase__ ): a : Any = i a : List[Any] = CharacterTokenizer(vocab=lowerCAmelCase__ , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("こんにちは" ) , ["こ", "ん", "に", "ち", "は"] ) self.assertListEqual(tokenizer.tokenize("こんにちほ" ) , ["こ", "ん", "に", "ち", "[UNK]"] ) def __a ( self ) -> List[str]: a : Dict = self.tokenizer_class.from_pretrained("cl-tohoku/bert-base-japanese-char" ) a : List[str] = tokenizer.encode("ありがとう。" , add_special_tokens=lowerCAmelCase__ ) a : Optional[int] = tokenizer.encode("どういたしまして。" , add_special_tokens=lowerCAmelCase__ ) a : Dict = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ) a : List[Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[str]: a : List[Any] = "cl-tohoku/bert-base-japanese" a : Any = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[str]: a : str = "cl-tohoku/bert-base-japanese" with self.assertLogs("transformers" , level="WARNING" ) as cm: BertTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) ) a : Union[str, Any] = "bert-base-cased" with self.assertLogs("transformers" , level="WARNING" ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertTrue( cm.records[0].message.startswith( "The tokenizer class you load from this checkpoint is not the same type as the class this function" " is called from." ) )
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __UpperCamelCase : lowerCamelCase : Optional[int] =BlenderbotConfig lowerCamelCase : Dict ={} lowerCamelCase : Any ="""gelu""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=2 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=20 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , ) -> Dict: a : List[str] = parent a : List[str] = batch_size a : Optional[Any] = seq_length a : Any = is_training a : Optional[int] = use_labels a : Any = vocab_size a : Tuple = hidden_size a : Tuple = num_hidden_layers a : str = num_attention_heads a : str = intermediate_size a : Union[str, Any] = hidden_dropout_prob a : int = attention_probs_dropout_prob a : Optional[Any] = max_position_embeddings a : Any = eos_token_id a : Optional[int] = pad_token_id a : Any = bos_token_id def __a ( self ) -> Any: a : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) a : Tuple = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) a : Any = tf.concat([input_ids, eos_tensor] , axis=1 ) a : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : str = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) a : Optional[Any] = prepare_blenderbot_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : Tuple = TFBlenderbotModel(config=lowerCAmelCase__ ).get_decoder() a : Any = inputs_dict["input_ids"] a : Tuple = input_ids[:1, :] a : Dict = inputs_dict["attention_mask"][:1, :] a : List[Any] = inputs_dict["head_mask"] a : Any = 1 # first forward pass a : Optional[Any] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , head_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) a, a : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids a : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size ) a : Any = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and a : int = tf.concat([input_ids, next_tokens] , axis=-1 ) a : List[Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) a : Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] a : List[Any] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice a : int = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) a : int = output_from_no_past[:, -3:, random_slice_idx] a : Any = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1E-3 ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : Any , _lowercase : Optional[Any] , _lowercase : Any=None , _lowercase : Union[str, Any]=None , _lowercase : Union[str, Any]=None , _lowercase : Tuple=None , _lowercase : Dict=None , ) ->Dict: '''simple docstring''' if attention_mask is None: a : str = tf.cast(tf.math.not_equal(_lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: a : Union[str, Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: a : int = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: a : Optional[int] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: a : List[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : str =(TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () lowerCamelCase : Optional[Any] =(TFBlenderbotForConditionalGeneration,) if is_tf_available() else () lowerCamelCase : Any =( { """conversational""": TFBlenderbotForConditionalGeneration, """feature-extraction""": TFBlenderbotModel, """summarization""": TFBlenderbotForConditionalGeneration, """text2text-generation""": TFBlenderbotForConditionalGeneration, """translation""": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) lowerCamelCase : Tuple =True lowerCamelCase : Optional[int] =False lowerCamelCase : List[Any] =False def __a ( self ) -> Optional[Any]: a : Union[str, Any] = TFBlenderbotModelTester(self ) a : Union[str, Any] = ConfigTester(self , config_class=lowerCAmelCase__ ) def __a ( self ) -> List[str]: self.config_tester.run_common_tests() def __a ( self ) -> int: a : Tuple = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase__ ) @require_tokenizers @require_tf class __UpperCamelCase ( unittest.TestCase ): lowerCamelCase : int =["""My friends are cool but they eat too many carbs."""] lowerCamelCase : List[Any] ="""facebook/blenderbot-400M-distill""" @cached_property def __a ( self ) -> Any: return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def __a ( self ) -> List[Any]: a : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def __a ( self ) -> Union[str, Any]: a : Optional[int] = self.tokenizer(self.src_text , return_tensors="tf" ) a : Union[str, Any] = self.model.generate( model_inputs.input_ids , ) a : Optional[int] = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowerCAmelCase__ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa a : List[Any] = logging.getLogger(__name__) class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[int] ="""summarization""" lowerCamelCase : int =["""loss"""] lowerCamelCase : Dict =ROUGE_KEYS lowerCamelCase : Optional[Any] ="""rouge2""" def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Tuple: if hparams.sortish_sampler and hparams.gpus > 1: a : Union[str, Any] = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError("Dynamic Batch size does not work for multi-gpu training" ) if hparams.sortish_sampler: raise ValueError("--sortish_sampler and --max_tokens_per_batch may not be used simultaneously" ) super().__init__(lowerCAmelCase__ , num_labels=lowerCAmelCase__ , mode=self.mode , **lowerCAmelCase__ ) use_task_specific_params(self.model , "summarization" ) save_git_info(self.hparams.output_dir ) a : Optional[Any] = Path(self.output_dir ) / "metrics.json" a : int = Path(self.output_dir ) / "hparams.pkl" pickle_save(self.hparams , self.hparams_save_path ) a : List[Any] = 0 a : Union[str, Any] = defaultdict(lowerCAmelCase__ ) a : Any = self.config.model_type a : str = self.config.tgt_vocab_size if self.model_type == "fsmt" else self.config.vocab_size a : dict = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } a : Tuple = { "train": self.hparams.n_train, "val": self.hparams.n_val, "test": self.hparams.n_test, } a : Optional[Any] = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} a : str = { "train": self.hparams.max_target_length, "val": self.hparams.val_max_target_length, "test": self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], f"""target_lens: {self.target_lens}""" assert self.target_lens["train"] <= self.target_lens["test"], f"""target_lens: {self.target_lens}""" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) a : Optional[Any] = get_git_info()["repo_sha"] a : List[Any] = hparams.num_workers a : Union[str, Any] = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , lowerCAmelCase__ ): a : Optional[int] = self.tokenizer.lang_code_to_id[hparams.tgt_lang] a : Optional[Any] = self.decoder_start_token_id a : Union[str, Any] = ( SeqaSeqDataset if hasattr(self.tokenizer , "prepare_seq2seq_batch" ) else LegacySeqaSeqDataset ) a : Optional[Any] = False a : int = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: a : Any = self.hparams.eval_max_gen_length else: a : Optional[int] = self.model.config.max_length a : Optional[Any] = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def __a ( self , lowerCAmelCase__ ) -> Dict[str, List[str]]: a : str = { k: self.tokenizer.batch_decode(v.tolist() ) if "mask" not in k else v.shape for k, v in batch.items() } save_json(lowerCAmelCase__ , Path(self.output_dir ) / "text_batch.json" ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / "tok_batch.json" ) a : Any = True return readable_batch def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: return self.model(lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> int: a : str = self.tokenizer.batch_decode( lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ ) return lmap(str.strip , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> Tuple: a : List[Any] = self.tokenizer.pad_token_id a, a : Optional[int] = batch["input_ids"], batch["attention_mask"] a : Tuple = batch["labels"] if isinstance(self.model , lowerCAmelCase__ ): a : Optional[int] = self.model._shift_right(lowerCAmelCase__ ) else: a : Optional[int] = shift_tokens_right(lowerCAmelCase__ , lowerCAmelCase__ ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero a : List[str] = decoder_input_ids self.save_readable_batch(lowerCAmelCase__ ) a : Any = self(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , decoder_input_ids=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) a : Any = outputs["logits"] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id a : List[str] = nn.CrossEntropyLoss(ignore_index=lowerCAmelCase__ ) assert lm_logits.shape[-1] == self.vocab_size a : Optional[int] = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: a : Tuple = nn.functional.log_softmax(lowerCAmelCase__ , dim=-1 ) a, a : Any = label_smoothed_nll_loss( lowerCAmelCase__ , lowerCAmelCase__ , self.hparams.label_smoothing , ignore_index=lowerCAmelCase__ ) return (loss,) @property def __a ( self ) -> int: return self.tokenizer.pad_token_id def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: a : Optional[int] = self._step(lowerCAmelCase__ ) a : Dict = dict(zip(self.loss_names , lowerCAmelCase__ ) ) # tokens per batch a : List[Any] = batch["input_ids"].ne(self.pad ).sum() + batch["labels"].ne(self.pad ).sum() a : List[Any] = batch["input_ids"].shape[0] a : Any = batch["input_ids"].eq(self.pad ).sum() a : str = batch["input_ids"].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: return self._generative_step(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__="val" ) -> Dict: self.step_count += 1 a : List[str] = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} a : int = losses["loss"] a : Any = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ["gen_time", "gen_len"] } a : Optional[Any] = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) a : torch.FloatTensor = torch.tensor(lowerCAmelCase__ ).type_as(lowerCAmelCase__ ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(lowerCAmelCase__ ) a : Optional[int] = {f"""{prefix}_avg_{k}""": x for k, x in losses.items()} a : List[Any] = self.step_count self.metrics[prefix].append(lowerCAmelCase__ ) # callback writes this to self.metrics_save_path a : Optional[int] = flatten_list([x["preds"] for x in outputs] ) return { "log": all_metrics, "preds": preds, f"""{prefix}_loss""": loss, f"""{prefix}_{self.val_metric}""": metric_tensor, } def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: return calculate_rouge(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> dict: a : Any = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') a : Optional[int] = self.model.generate( batch["input_ids"] , attention_mask=batch["attention_mask"] , use_cache=lowerCAmelCase__ , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) a : str = (time.time() - ta) / batch["input_ids"].shape[0] a : List[str] = self.ids_to_clean_text(lowerCAmelCase__ ) a : List[str] = self.ids_to_clean_text(batch["labels"] ) a : int = self._step(lowerCAmelCase__ ) a : int = dict(zip(self.loss_names , lowerCAmelCase__ ) ) a : Dict = self.calc_generative_metrics(lowerCAmelCase__ , lowerCAmelCase__ ) a : Optional[int] = np.mean(lmap(lowerCAmelCase__ , lowerCAmelCase__ ) ) base_metrics.update(gen_time=lowerCAmelCase__ , gen_len=lowerCAmelCase__ , preds=lowerCAmelCase__ , target=lowerCAmelCase__ , **lowerCAmelCase__ ) return base_metrics def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: return self._generative_step(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: return self.validation_epoch_end(lowerCAmelCase__ , prefix="test" ) def __a ( self , lowerCAmelCase__ ) -> SeqaSeqDataset: a : Union[str, Any] = self.n_obs[type_path] a : List[str] = self.target_lens[type_path] a : Any = self.dataset_class( self.tokenizer , type_path=lowerCAmelCase__ , n_obs=lowerCAmelCase__ , max_target_length=lowerCAmelCase__ , **self.dataset_kwargs , ) return dataset def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> DataLoader: a : Tuple = self.get_dataset(lowerCAmelCase__ ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": a : int = dataset.make_sortish_sampler(lowerCAmelCase__ , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , collate_fn=dataset.collate_fn , shuffle=lowerCAmelCase__ , num_workers=self.num_workers , sampler=lowerCAmelCase__ , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": a : Union[str, Any] = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCAmelCase__ , batch_sampler=lowerCAmelCase__ , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , collate_fn=dataset.collate_fn , shuffle=lowerCAmelCase__ , num_workers=self.num_workers , sampler=lowerCAmelCase__ , ) def __a ( self ) -> DataLoader: a : List[str] = self.get_dataloader("train" , batch_size=self.hparams.train_batch_size , shuffle=lowerCAmelCase__ ) return dataloader def __a ( self ) -> DataLoader: return self.get_dataloader("val" , batch_size=self.hparams.eval_batch_size ) def __a ( self ) -> DataLoader: return self.get_dataloader("test" , batch_size=self.hparams.eval_batch_size ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: BaseTransformer.add_model_specific_args(lowerCAmelCase__ , lowerCAmelCase__ ) add_generic_args(lowerCAmelCase__ , lowerCAmelCase__ ) parser.add_argument( "--max_source_length" , default=1024 , type=lowerCAmelCase__ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--max_target_length" , default=56 , type=lowerCAmelCase__ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--val_max_target_length" , default=142 , type=lowerCAmelCase__ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--test_max_target_length" , default=142 , type=lowerCAmelCase__ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument("--freeze_encoder" , action="store_true" ) parser.add_argument("--freeze_embeds" , action="store_true" ) parser.add_argument("--sortish_sampler" , action="store_true" , default=lowerCAmelCase__ ) parser.add_argument("--overwrite_output_dir" , action="store_true" , default=lowerCAmelCase__ ) parser.add_argument("--max_tokens_per_batch" , type=lowerCAmelCase__ , default=lowerCAmelCase__ ) parser.add_argument("--logger_name" , type=lowerCAmelCase__ , choices=["default", "wandb", "wandb_shared"] , default="default" ) parser.add_argument("--n_train" , type=lowerCAmelCase__ , default=-1 , required=lowerCAmelCase__ , help="# examples. -1 means use all." ) parser.add_argument("--n_val" , type=lowerCAmelCase__ , default=500 , required=lowerCAmelCase__ , help="# examples. -1 means use all." ) parser.add_argument("--n_test" , type=lowerCAmelCase__ , default=-1 , required=lowerCAmelCase__ , help="# examples. -1 means use all." ) parser.add_argument( "--task" , type=lowerCAmelCase__ , default="summarization" , required=lowerCAmelCase__ , help="# examples. -1 means use all." ) parser.add_argument("--label_smoothing" , type=lowerCAmelCase__ , default=0.0 , required=lowerCAmelCase__ ) parser.add_argument("--src_lang" , type=lowerCAmelCase__ , default="" , required=lowerCAmelCase__ ) parser.add_argument("--tgt_lang" , type=lowerCAmelCase__ , default="" , required=lowerCAmelCase__ ) parser.add_argument("--eval_beams" , type=lowerCAmelCase__ , default=lowerCAmelCase__ , required=lowerCAmelCase__ ) parser.add_argument( "--val_metric" , type=lowerCAmelCase__ , default=lowerCAmelCase__ , required=lowerCAmelCase__ , choices=["bleu", "rouge2", "loss", None] ) parser.add_argument("--eval_max_gen_length" , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help="never generate more than n tokens" ) parser.add_argument("--save_top_k" , type=lowerCAmelCase__ , default=1 , required=lowerCAmelCase__ , help="How many checkpoints to save" ) parser.add_argument( "--early_stopping_patience" , type=lowerCAmelCase__ , default=-1 , required=lowerCAmelCase__ , help=( "-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So" " val_check_interval will effect it." ) , ) return parser class __UpperCamelCase ( a__ ): lowerCamelCase : Tuple ="""translation""" lowerCamelCase : Optional[int] =["""loss"""] lowerCamelCase : Tuple =["""bleu"""] lowerCamelCase : Tuple ="""bleu""" def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[Any]: super().__init__(lowerCAmelCase__ , **lowerCAmelCase__ ) a : Optional[Any] = hparams.src_lang a : Union[str, Any] = hparams.tgt_lang def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> dict: return calculate_bleu(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Tuple=None ) ->SummarizationModule: '''simple docstring''' Path(args.output_dir ).mkdir(exist_ok=_lowercase ) check_output_dir(_lowercase , expected_items=3 ) if model is None: if "summarization" in args.task: a : SummarizationModule = SummarizationModule(_lowercase ) else: a : SummarizationModule = TranslationModule(_lowercase ) a : int = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith("/tmp" ) or str(args.output_dir ).startswith("/var" ) ): a : Tuple = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger a : Optional[int] = os.environ.get("WANDB_PROJECT" , _lowercase ) a : Dict = WandbLogger(name=model.output_dir.name , project=_lowercase ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger a : str = WandbLogger(name=model.output_dir.name , project=F"""hf_{dataset}""" ) if args.early_stopping_patience >= 0: a : Tuple = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: a : Optional[int] = False a : Optional[Any] = args.val_metric == "loss" a : pl.Trainer = generic_train( _lowercase , _lowercase , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , _lowercase ) , early_stopping_callback=_lowercase , logger=_lowercase , ) pickle_save(model.hparams , model.output_dir / "hparams.pkl" ) if not args.do_predict: return model a : Optional[Any] = "" a : Dict = sorted(glob.glob(os.path.join(args.output_dir , "*.ckpt" ) , recursive=_lowercase ) ) if checkpoints: a : Optional[int] = checkpoints[-1] a : int = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": a : List[str] = argparse.ArgumentParser() a : Optional[int] = pl.Trainer.add_argparse_args(parser) a : int = SummarizationModule.add_model_specific_args(parser, os.getcwd()) a : Union[str, Any] = parser.parse_args() main(args)
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version a : List[str] = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize a : Tuple = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' a : Optional[Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' a : str = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def __a ( self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=0.9 , lowerCAmelCase__=3 , lowerCAmelCase__=0.5 ) -> int: if NLTK_VERSION >= version.Version("3.6.5" ): a : List[str] = [ meteor_score.single_meteor_score( word_tokenize(lowerCAmelCase__ ) , word_tokenize(lowerCAmelCase__ ) , alpha=lowerCAmelCase__ , beta=lowerCAmelCase__ , gamma=lowerCAmelCase__ ) for ref, pred in zip(lowerCAmelCase__ , lowerCAmelCase__ ) ] else: a : List[Any] = [ meteor_score.single_meteor_score(lowerCAmelCase__ , lowerCAmelCase__ , alpha=lowerCAmelCase__ , beta=lowerCAmelCase__ , gamma=lowerCAmelCase__ ) for ref, pred in zip(lowerCAmelCase__ , lowerCAmelCase__ ) ] return {"meteor": np.mean(lowerCAmelCase__ )}
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Dict = logging.get_logger(__name__) a : int = '''▁''' a : Dict = {'''vocab_file''': '''vocab.txt''', '''sentencepiece_model_ckpt''': '''sentencepiece.bpe.model'''} a : Tuple = { '''sentencepiece_model_file''': '''sentencepiece.bpe.model''', '''vocab_file''': '''vocab.txt''', } a : Tuple = { '''vocab_file''': { '''ernie-m-base''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt''', '''ernie-m-large''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt''', }, '''sentencepiece_model_file''': { '''ernie-m-base''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model''', '''ernie-m-large''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model''', }, } a : Any = { '''ernie-m-base''': 514, '''ernie-m-large''': 514, } a : Dict = { '''ernie-m-base''': {'''do_lower_case''': False}, '''ernie-m-large''': {'''do_lower_case''': False}, } class __UpperCamelCase ( a__ ): lowerCamelCase : List[str] =["input_ids"] lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_INIT_CONFIGURATION lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Dict =RESOURCE_FILES_NAMES def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__="utf8" , lowerCAmelCase__="[UNK]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="[PAD]" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , vocab_file=lowerCAmelCase__ , encoding=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase__ , ) a : List[str] = do_lower_case a : Any = sentencepiece_model_ckpt a : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase__ ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: a : Any = self.load_vocab(filepath=lowerCAmelCase__ ) else: a : Dict = {self.sp_model.id_to_piece(lowerCAmelCase__ ): id for id in range(self.sp_model.get_piece_size() )} a : Any = {v: k for k, v in self.vocab.items()} def __a ( self , lowerCAmelCase__ ) -> Any: if text is None: return None a : List[Any] = self.tokenize(lowerCAmelCase__ ) a, a : Any = "", [] for i, ch in enumerate(lowerCAmelCase__ ): if ch in self.SP_CHAR_MAPPING: a : Any = self.SP_CHAR_MAPPING.get(lowerCAmelCase__ ) else: a : Optional[Any] = unicodedata.normalize("NFKC" , lowerCAmelCase__ ) if self.is_whitespace(lowerCAmelCase__ ): continue normalized_text += ch char_mapping.extend([i] * len(lowerCAmelCase__ ) ) a, a, a : Union[str, Any] = normalized_text, [], 0 if self.do_lower_case: a : List[Any] = text.lower() for token in split_tokens: if token[:1] == "▁": a : Tuple = token[1:] a : Any = text[offset:].index(lowerCAmelCase__ ) + offset a : Union[str, Any] = start + len(lowerCAmelCase__ ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) a : Any = end return token_mapping @property def __a ( self ) -> Dict: return len(self.vocab ) def __a ( self ) -> List[Any]: return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self ) -> Optional[Any]: a : Optional[int] = self.__dict__.copy() a : Any = None return state def __setstate__( self , lowerCAmelCase__ ) -> Optional[int]: a : str = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): a : int = {} a : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def __a ( self , lowerCAmelCase__ ) -> List[str]: return "".join((self.SP_CHAR_MAPPING.get(lowerCAmelCase__ , lowerCAmelCase__ ) for c in text) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=64 , lowerCAmelCase__=0.1 ) -> List[Any]: if self.sp_model_kwargs.get("enable_sampling" ) is True: a : int = True if self.sp_model_kwargs.get("alpha" ) is not None: a : Any = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: a : str = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: a : int = self.sp_model.EncodeAsPieces(lowerCAmelCase__ ) else: a : List[str] = self.sp_model.SampleEncodeAsPieces(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : Optional[Any] = [] for pi, piece in enumerate(lowerCAmelCase__ ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(lowerCAmelCase__ ) and pi != 0: new_pieces.append(lowerCAmelCase__ ) continue else: continue a : Union[str, Any] = 0 for i, chunk in enumerate(lowerCAmelCase__ ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(lowerCAmelCase__ ) or self.is_punct(lowerCAmelCase__ ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(lowerCAmelCase__ ) a : Tuple = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) a : Dict = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) a : Optional[Any] = i if len(lowerCAmelCase__ ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: a : Any = "".join(lowerCAmelCase__ ).replace(lowerCAmelCase__ , " " ).strip() return out_string def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : Optional[Any] = self.convert_ids_to_tokens(lowerCAmelCase__ ) a : Optional[Any] = "".join(lowerCAmelCase__ ).replace(lowerCAmelCase__ , " " ).strip() return out_string def __a ( self , lowerCAmelCase__ ) -> str: return self.vocab.get(lowerCAmelCase__ , self.vocab.get(self.unk_token ) ) def __a ( self , lowerCAmelCase__ ) -> str: return self.reverse_vocab.get(lowerCAmelCase__ , self.unk_token ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> Dict: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a : Tuple = [self.cls_token_id] a : Any = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> Any: if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=False ) -> Tuple: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: # called when `add_special_tokens` is True, so align with `build_inputs_with_special_tokens` method if token_ids_a is None: # [CLS] X [SEP] return (len(lowerCAmelCase__ ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(lowerCAmelCase__ ) + 1) + [1] * (len(lowerCAmelCase__ ) + 3) def __a ( self , lowerCAmelCase__ ) -> List[str]: if "\u4e00" <= char <= "\u9fff": return True return False def __a ( self , lowerCAmelCase__ ) -> str: if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def __a ( self , lowerCAmelCase__ ) -> Any: if char in ",;:.?!~,;:。?!《》【】": return True return False def __a ( self , lowerCAmelCase__ ) -> Dict: if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(lowerCAmelCase__ ) == 1: a : List[str] = unicodedata.category(lowerCAmelCase__ ) if cat == "Zs": return True return False def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : List[Any] = {} with io.open(lowerCAmelCase__ , "r" , encoding="utf-8" ) as f: for index, line in enumerate(lowerCAmelCase__ ): a : List[Any] = line.rstrip("\n" ) a : List[Any] = int(lowerCAmelCase__ ) return token_to_idx def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: a : List[str] = 0 if os.path.isdir(lowerCAmelCase__ ): a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: a : Optional[int] = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" " Please check that the vocabulary is not corrupted!" ) a : str = token_index writer.write(token + "\n" ) index += 1 a : Any = os.path.join(lowerCAmelCase__ , "sentencepiece.bpe.model" ) with open(lowerCAmelCase__ , "wb" ) as fi: a : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__ ) return (vocab_file,)
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" from math import isqrt def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->bool: '''simple docstring''' return all(number % divisor != 0 for divisor in range(2 , isqrt(_lowercase ) + 1 ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10**6 ) ->int: '''simple docstring''' a : Optional[Any] = 0 a : int = 1 a : Dict = 7 while prime_candidate < max_prime: primes_count += is_prime(_lowercase ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F'''{solution() = }''')
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class __UpperCamelCase : def __init__( self , lowerCAmelCase__=2 , lowerCAmelCase__=3 , lowerCAmelCase__=64 , lowerCAmelCase__=None ) -> Optional[int]: a : Optional[Any] = np.random.default_rng(lowerCAmelCase__ ) a : str = length a : List[str] = rng.normal(size=(length,) ).astype(np.floataa ) a : Union[str, Any] = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> Any: return self.length def __getitem__( self , lowerCAmelCase__ ) -> str: return {"x": self.x[i], "y": self.y[i]} class __UpperCamelCase ( torch.nn.Module ): def __init__( self , lowerCAmelCase__=0 , lowerCAmelCase__=0 , lowerCAmelCase__=False ) -> str: super().__init__() a : int = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) a : Optional[Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) a : List[str] = True def __a ( self , lowerCAmelCase__=None ) -> List[str]: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) a : Tuple = False return x * self.a[0] + self.b[0] class __UpperCamelCase ( torch.nn.Module ): def __init__( self , lowerCAmelCase__=0 , lowerCAmelCase__=0 , lowerCAmelCase__=False ) -> int: super().__init__() a : Tuple = torch.nn.Parameter(torch.tensor(lowerCAmelCase__ ).float() ) a : Any = torch.nn.Parameter(torch.tensor(lowerCAmelCase__ ).float() ) a : Optional[int] = True def __a ( self , lowerCAmelCase__=None ) -> Tuple: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) a : Optional[Any] = False return x * self.a + self.b def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int = 16 ) ->Dict: '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer a : Optional[int] = AutoTokenizer.from_pretrained("bert-base-cased" ) a : Optional[int] = {"train": "tests/test_samples/MRPC/train.csv", "validation": "tests/test_samples/MRPC/dev.csv"} a : str = load_dataset("csv" , data_files=_lowercase ) a : Dict = datasets["train"].unique("label" ) a : List[str] = {v: i for i, v in enumerate(_lowercase )} def tokenize_function(_lowercase : Optional[int] ): # max_length=None => use the model max length (it's actually the default) a : List[str] = tokenizer( examples["sentence1"] , examples["sentence2"] , truncation=_lowercase , max_length=_lowercase , padding="max_length" ) if "label" in examples: a : Optional[Any] = [label_to_id[l] for l in examples["label"]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset a : Any = datasets.map( _lowercase , batched=_lowercase , remove_columns=["sentence1", "sentence2", "label"] , ) def collate_fn(_lowercase : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(_lowercase , padding="max_length" , max_length=128 , return_tensors="pt" ) return tokenizer.pad(_lowercase , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. a : List[Any] = DataLoader(tokenized_datasets["train"] , shuffle=_lowercase , collate_fn=_lowercase , batch_size=2 ) a : Optional[Any] = DataLoader(tokenized_datasets["validation"] , shuffle=_lowercase , collate_fn=_lowercase , batch_size=1 ) return train_dataloader, eval_dataloader
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" a : Optional[Any] = [ '''Audio''', '''Array2D''', '''Array3D''', '''Array4D''', '''Array5D''', '''ClassLabel''', '''Features''', '''Sequence''', '''Value''', '''Image''', '''Translation''', '''TranslationVariableLanguages''', ] from .audio import Audio from .features import ArrayaD, ArrayaD, ArrayaD, ArrayaD, ClassLabel, Features, Sequence, Value from .image import Image from .translation import Translation, TranslationVariableLanguages
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import fire from utils import calculate_rouge, save_json def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : List[str] , _lowercase : Tuple=None , **_lowercase : Optional[Any] ) ->List[Any]: '''simple docstring''' a : List[str] = [x.strip() for x in open(_lowercase ).readlines()] a : Any = [x.strip() for x in open(_lowercase ).readlines()][: len(_lowercase )] a : Dict = calculate_rouge(_lowercase , _lowercase , **_lowercase ) if save_path is not None: save_json(_lowercase , _lowercase , indent=_lowercase ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig a : Dict = logging.get_logger(__name__) # General docstring a : Optional[Any] = '''MobileNetV1Config''' # Base docstring a : Tuple = '''google/mobilenet_v1_1.0_224''' a : Dict = [1, 1024, 7, 7] # Image classification docstring a : str = '''google/mobilenet_v1_1.0_224''' a : Any = '''tabby, tabby cat''' a : List[Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Optional[Any] , _lowercase : Dict=None ) ->int: '''simple docstring''' a : Union[str, Any] = {} if isinstance(_lowercase , _lowercase ): a : List[Any] = model.mobilenet_va else: a : List[str] = model a : int = "MobilenetV1/Conv2d_0/" a : str = backbone.conv_stem.convolution.weight a : int = backbone.conv_stem.normalization.bias a : Optional[int] = backbone.conv_stem.normalization.weight a : List[Any] = backbone.conv_stem.normalization.running_mean a : List[str] = backbone.conv_stem.normalization.running_var for i in range(13 ): a : Dict = i + 1 a : Union[str, Any] = i * 2 a : Union[str, Any] = backbone.layer[pt_index] a : Optional[int] = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" a : str = pointer.convolution.weight a : str = pointer.normalization.bias a : Optional[Any] = pointer.normalization.weight a : List[str] = pointer.normalization.running_mean a : Optional[int] = pointer.normalization.running_var a : str = backbone.layer[pt_index + 1] a : List[str] = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" a : Tuple = pointer.convolution.weight a : Tuple = pointer.normalization.bias a : int = pointer.normalization.weight a : Dict = pointer.normalization.running_mean a : List[Any] = pointer.normalization.running_var if isinstance(_lowercase , _lowercase ): a : List[Any] = "MobilenetV1/Logits/Conv2d_1c_1x1/" a : int = model.classifier.weight a : Dict = model.classifier.bias return tf_to_pt_map def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int , _lowercase : Tuple ) ->List[Any]: '''simple docstring''' try: import numpy as np import tensorflow as tf except ImportError: logger.error( "Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions." ) raise # Load weights from TF model a : Any = tf.train.list_variables(_lowercase ) a : Tuple = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) a : int = tf.train.load_variable(_lowercase , _lowercase ) a : Union[str, Any] = array # Build TF to PyTorch weights loading map a : List[str] = _build_tf_to_pytorch_map(_lowercase , _lowercase , _lowercase ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue a : int = tf_weights[name] if "depthwise_weights" in name: logger.info("Transposing depthwise" ) a : List[Any] = np.transpose(_lowercase , (2, 3, 0, 1) ) elif "weights" in name: logger.info("Transposing" ) if len(pointer.shape ) == 2: # copying into linear layer a : List[str] = array.squeeze().transpose() else: a : Dict = np.transpose(_lowercase , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) a : List[Any] = torch.from_numpy(_lowercase ) tf_weights.pop(_lowercase , _lowercase ) tf_weights.pop(name + "/RMSProp" , _lowercase ) tf_weights.pop(name + "/RMSProp_1" , _lowercase ) tf_weights.pop(name + "/ExponentialMovingAverage" , _lowercase ) logger.info(F"""Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}""" ) return model def _SCREAMING_SNAKE_CASE ( _lowercase : torch.Tensor , _lowercase : nn.Convad ) ->torch.Tensor: '''simple docstring''' a, a : Tuple = features.shape[-2:] a, a : List[Any] = conv_layer.stride a, a : Any = conv_layer.kernel_size if in_height % stride_height == 0: a : Union[str, Any] = max(kernel_height - stride_height , 0 ) else: a : int = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: a : Union[str, Any] = max(kernel_width - stride_width , 0 ) else: a : Optional[Any] = max(kernel_width - (in_width % stride_width) , 0 ) a : Tuple = pad_along_width // 2 a : Union[str, Any] = pad_along_width - pad_left a : str = pad_along_height // 2 a : Tuple = pad_along_height - pad_top a : Any = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(_lowercase , _lowercase , "constant" , 0.0 ) class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = False , lowerCAmelCase__ = True , lowerCAmelCase__ = True , ) -> None: super().__init__() a : str = config if in_channels % groups != 0: raise ValueError(f"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(f"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) a : List[str] = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) a : List[Any] = nn.Convad( in_channels=lowerCAmelCase__ , out_channels=lowerCAmelCase__ , kernel_size=lowerCAmelCase__ , stride=lowerCAmelCase__ , padding=lowerCAmelCase__ , groups=lowerCAmelCase__ , bias=lowerCAmelCase__ , padding_mode="zeros" , ) if use_normalization: a : Optional[int] = nn.BatchNormad( num_features=lowerCAmelCase__ , eps=config.layer_norm_eps , momentum=0.9_997 , affine=lowerCAmelCase__ , track_running_stats=lowerCAmelCase__ , ) else: a : List[str] = None if use_activation: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = ACTaFN[use_activation] elif isinstance(config.hidden_act , lowerCAmelCase__ ): a : Optional[Any] = ACTaFN[config.hidden_act] else: a : Any = config.hidden_act else: a : Tuple = None def __a ( self , lowerCAmelCase__ ) -> torch.Tensor: if self.config.tf_padding: a : str = apply_tf_padding(lowerCAmelCase__ , self.convolution ) a : Any = self.convolution(lowerCAmelCase__ ) if self.normalization is not None: a : str = self.normalization(lowerCAmelCase__ ) if self.activation is not None: a : Any = self.activation(lowerCAmelCase__ ) return features class __UpperCamelCase ( a__ ): lowerCamelCase : Any =MobileNetVaConfig lowerCamelCase : Any =load_tf_weights_in_mobilenet_va lowerCamelCase : Tuple ="""mobilenet_v1""" lowerCamelCase : List[Any] ="""pixel_values""" lowerCamelCase : List[str] =False def __a ( self , lowerCAmelCase__ ) -> None: if isinstance(lowerCAmelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(lowerCAmelCase__ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) a : Any = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' a : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.""" , a__ , ) class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = True ) -> int: super().__init__(lowerCAmelCase__ ) a : Any = config a : Tuple = 32 a : Union[str, Any] = max(int(depth * config.depth_multiplier ) , config.min_depth ) a : Optional[Any] = MobileNetVaConvLayer( lowerCAmelCase__ , in_channels=config.num_channels , out_channels=lowerCAmelCase__ , kernel_size=3 , stride=2 , ) a : Optional[int] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] a : List[str] = nn.ModuleList() for i in range(13 ): a : Union[str, Any] = out_channels if strides[i] == 2 or i == 0: depth *= 2 a : int = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( lowerCAmelCase__ , in_channels=lowerCAmelCase__ , out_channels=lowerCAmelCase__ , kernel_size=3 , stride=strides[i] , groups=lowerCAmelCase__ , ) ) self.layer.append( MobileNetVaConvLayer( lowerCAmelCase__ , in_channels=lowerCAmelCase__ , out_channels=lowerCAmelCase__ , kernel_size=1 , ) ) a : Any = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def __a ( self , lowerCAmelCase__ ) -> str: raise NotImplementedError @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __a ( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: a : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : List[Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) a : str = self.conv_stem(lowerCAmelCase__ ) a : Tuple = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): a : Optional[int] = layer_module(lowerCAmelCase__ ) if output_hidden_states: a : List[str] = all_hidden_states + (hidden_states,) a : List[Any] = hidden_states if self.pooler is not None: a : str = torch.flatten(self.pooler(lowerCAmelCase__ ) , start_dim=1 ) else: a : str = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCAmelCase__ , pooler_output=lowerCAmelCase__ , hidden_states=lowerCAmelCase__ , ) @add_start_docstrings( """ MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , a__ , ) class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ ) -> None: super().__init__(lowerCAmelCase__ ) a : Tuple = config.num_labels a : Optional[Any] = MobileNetVaModel(lowerCAmelCase__ ) a : List[Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head a : str = nn.Dropout(config.classifier_dropout_prob , inplace=lowerCAmelCase__ ) a : int = nn.Linear(lowerCAmelCase__ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __a ( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: a : Tuple = return_dict if return_dict is not None else self.config.use_return_dict a : Dict = self.mobilenet_va(lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ ) a : Tuple = outputs.pooler_output if return_dict else outputs[1] a : Optional[int] = self.classifier(self.dropout(lowerCAmelCase__ ) ) a : Optional[Any] = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: a : List[Any] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): a : Optional[Any] = "single_label_classification" else: a : Optional[int] = "multi_label_classification" if self.config.problem_type == "regression": a : Any = MSELoss() if self.num_labels == 1: a : Optional[Any] = loss_fct(logits.squeeze() , labels.squeeze() ) else: a : Dict = loss_fct(lowerCAmelCase__ , lowerCAmelCase__ ) elif self.config.problem_type == "single_label_classification": a : Any = CrossEntropyLoss() a : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": a : Union[str, Any] = BCEWithLogitsLoss() a : Optional[Any] = loss_fct(lowerCAmelCase__ , lowerCAmelCase__ ) if not return_dict: a : Optional[int] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=lowerCAmelCase__ , logits=lowerCAmelCase__ , hidden_states=outputs.hidden_states , )
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class __UpperCamelCase ( unittest.TestCase , a__ ): def __a ( self ) -> Union[str, Any]: a : int = load_tool("text-to-speech" ) self.tool.setup() def __a ( self ) -> Optional[int]: # SpeechT5 isn't deterministic torch.manual_seed(0 ) a : List[str] = self.tool("hey" ) a : Tuple = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_005_966_668_832_115_829, -0.0_003_657_640_190_795_064, -0.00_013_439_502_799_883_485] ) , ) ) def __a ( self ) -> List[str]: # SpeechT5 isn't deterministic torch.manual_seed(0 ) a : List[str] = self.tool("hey" ) a : int = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_005_966_668_832_115_829, -0.0_003_657_640_190_795_064, -0.00_013_439_502_799_883_485] ) , ) )
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" from collections.abc import Sequence from queue import Queue class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Optional[int]: a : List[str] = start a : Dict = end a : Optional[int] = val a : List[str] = (start + end) // 2 a : Any = left a : str = right def __repr__( self ) -> List[Any]: return f"""SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})""" class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : Tuple = collection a : Optional[int] = function if self.collection: a : int = self._build_tree(0 , len(lowerCAmelCase__ ) - 1 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: self._update_tree(self.root , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: return self._query_range(self.root , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: if start == end: return SegmentTreeNode(lowerCAmelCase__ , lowerCAmelCase__ , self.collection[start] ) a : int = (start + end) // 2 a : List[str] = self._build_tree(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = self._build_tree(mid + 1 , lowerCAmelCase__ ) return SegmentTreeNode(lowerCAmelCase__ , lowerCAmelCase__ , self.fn(left.val , right.val ) , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: if node.start == i and node.end == i: a : str = val return if i <= node.mid: self._update_tree(node.left , lowerCAmelCase__ , lowerCAmelCase__ ) else: self._update_tree(node.right , lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.fn(node.left.val , node.right.val ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , lowerCAmelCase__ , lowerCAmelCase__ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , lowerCAmelCase__ , node.mid ) , self._query_range(node.right , node.mid + 1 , lowerCAmelCase__ ) , ) else: # range in right child tree return self._query_range(node.right , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Any: if self.root is not None: a : Union[str, Any] = Queue() queue.put(self.root ) while not queue.empty(): a : Any = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 50) a : List[Any] = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase a : List[Any] = logging.get_logger(__name__) a : Tuple = { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json''', '''allenai/longformer-large-4096''': '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json''', '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json''' ), } class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] ="""longformer""" def __init__( self , lowerCAmelCase__ = 512 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 0 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 3_0522 , lowerCAmelCase__ = 768 , lowerCAmelCase__ = 12 , lowerCAmelCase__ = 12 , lowerCAmelCase__ = 3072 , lowerCAmelCase__ = "gelu" , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 512 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 1E-12 , lowerCAmelCase__ = False , **lowerCAmelCase__ , ) -> int: super().__init__(pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Optional[Any] = attention_window a : Dict = sep_token_id a : int = bos_token_id a : Union[str, Any] = eos_token_id a : Tuple = vocab_size a : Optional[Any] = hidden_size a : Union[str, Any] = num_hidden_layers a : int = num_attention_heads a : List[Any] = hidden_act a : Tuple = intermediate_size a : int = hidden_dropout_prob a : Tuple = attention_probs_dropout_prob a : Any = max_position_embeddings a : Union[str, Any] = type_vocab_size a : Dict = initializer_range a : str = layer_norm_eps a : Tuple = onnx_export class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = "default" , lowerCAmelCase__ = None ) -> Dict: super().__init__(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = True @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": a : Any = {0: "batch", 1: "choice", 2: "sequence"} else: a : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("global_attention_mask", dynamic_axis), ] ) @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: a : Tuple = super().outputs if self.task == "default": a : Any = {0: "batch"} return outputs @property def __a ( self ) -> float: return 1E-4 @property def __a ( self ) -> int: # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , ) -> Mapping[str, Any]: a : Tuple = super().generate_dummy_inputs( preprocessor=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , seq_length=lowerCAmelCase__ , is_pair=lowerCAmelCase__ , framework=lowerCAmelCase__ ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly a : Optional[int] = torch.zeros_like(inputs["input_ids"] ) # make every second token global a : Optional[Any] = 1 return inputs
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser( description=( '''Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned''' ''' Distillation''' ) ) parser.add_argument('''--model_type''', default='''roberta''', choices=['''roberta''', '''gpt2''']) parser.add_argument('''--model_name''', default='''roberta-large''', type=str) parser.add_argument('''--dump_checkpoint''', default='''serialization_dir/tf_roberta_048131723.pth''', type=str) parser.add_argument('''--vocab_transform''', action='''store_true''') a : str = parser.parse_args() if args.model_type == "roberta": a : int = RobertaForMaskedLM.from_pretrained(args.model_name) a : Dict = '''roberta''' elif args.model_type == "gpt2": a : Tuple = GPTaLMHeadModel.from_pretrained(args.model_name) a : List[str] = '''transformer''' a : Dict = model.state_dict() a : int = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: a : List[str] = state_dict[F'''{prefix}.{param_name}'''] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: a : List[Any] = F'''{prefix}.embeddings.{w}.weight''' a : Optional[Any] = state_dict[param_name] for w in ["weight", "bias"]: a : Tuple = F'''{prefix}.embeddings.LayerNorm.{w}''' a : Dict = state_dict[param_name] # Transformer Blocks # a : List[str] = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: a : int = state_dict[ F'''{prefix}.h.{teacher_idx}.{layer}.{w}''' ] a : Union[str, Any] = state_dict[F'''{prefix}.h.{teacher_idx}.attn.bias'''] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: a : int = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}''' ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: a : Tuple = state_dict[F'''{layer}'''] if args.vocab_transform: for w in ["weight", "bias"]: a : Tuple = state_dict[F'''lm_head.dense.{w}'''] a : Tuple = state_dict[F'''lm_head.layer_norm.{w}'''] elif args.model_type == "gpt2": for w in ["weight", "bias"]: a : Any = state_dict[F'''{prefix}.ln_f.{w}'''] a : Union[str, Any] = state_dict['''lm_head.weight'''] print(F'''N layers selected for distillation: {std_idx}''') print(F'''Number of params transferred for distillation: {len(compressed_sd.keys())}''') print(F'''Save transferred checkpoint to {args.dump_checkpoint}.''') torch.save(compressed_sd, args.dump_checkpoint)
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int ) ->list: '''simple docstring''' a : Dict = word.split() def justify(_lowercase : list , _lowercase : int , _lowercase : int ) -> str: a : Optional[Any] = max_width - width a : Tuple = len(_lowercase ) if len(_lowercase ) == 1: # if there is only word in line # just insert overall_spaces_count for the remainder of line return line[0] + " " * overall_spaces_count else: a : Any = words_count - 1 # num_spaces_between_words_list[i] : tells you to insert # num_spaces_between_words_list[i] spaces # after word on line[i] a : List[str] = spaces_to_insert_between_words * [ overall_spaces_count // spaces_to_insert_between_words ] a : Optional[int] = ( overall_spaces_count % spaces_to_insert_between_words ) # distribute spaces via round robin to the left words for i in range(_lowercase ): num_spaces_between_words_list[i] += 1 a : Union[str, Any] = [] for i in range(_lowercase ): # add the word aligned_words_list.append(line[i] ) # add the spaces to insert aligned_words_list.append(num_spaces_between_words_list[i] * " " ) # just add the last word to the sentence aligned_words_list.append(line[-1] ) # join the aligned words list to form a justified line return "".join(_lowercase ) a : List[Any] = [] a : list[str] = [] a : Optional[int] = 0 for word in words: if width + len(_lowercase ) + len(_lowercase ) <= max_width: # keep adding words until we can fill out max_width # width = sum of length of all words (without overall_spaces_count) # len(word) = length of current word # len(line) = number of overall_spaces_count to insert between words line.append(_lowercase ) width += len(_lowercase ) else: # justify the line and add it to result answer.append(justify(_lowercase , _lowercase , _lowercase ) ) # reset new line and new width a, a : Dict = [word], len(_lowercase ) a : Dict = max_width - width - len(_lowercase ) answer.append(" ".join(_lowercase ) + (remaining_spaces + 1) * " " ) return answer if __name__ == "__main__": from doctest import testmod testmod()
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] ) ->int: '''simple docstring''' a : Union[str, Any] = len(_lowercase ) // 2 # choose the middle 3 elements a : int = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=4 , ) -> int: a : str = parent a : List[str] = batch_size a : str = seq_length a : str = is_training a : Dict = use_attention_mask a : str = use_token_type_ids a : List[str] = use_labels a : Optional[int] = vocab_size a : Tuple = hidden_size a : Optional[int] = num_hidden_layers a : Tuple = num_attention_heads a : Any = intermediate_size a : str = hidden_act a : Any = hidden_dropout_prob a : Optional[Any] = attention_probs_dropout_prob a : Optional[int] = max_position_embeddings a : Tuple = type_vocab_size a : List[str] = type_sequence_label_size a : List[Any] = initializer_range a : str = num_choices def __a ( self ) -> List[str]: a : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : List[Any] = None if self.use_attention_mask: a : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) a : Dict = None if self.use_token_type_ids: a : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a : List[Any] = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __a ( self ) -> List[str]: a : Union[str, Any] = self.prepare_config_and_inputs() a, a, a, a : int = config_and_inputs a : Optional[Any] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Optional[int] =( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def __a ( self ) -> Dict: a : Dict = FlaxAlbertModelTester(self ) @slow def __a ( self ) -> int: for model_class_name in self.all_model_classes: a : Optional[int] = model_class_name.from_pretrained("albert-base-v2" ) a : Tuple = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase__ ) @require_flax class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> List[Any]: a : Any = FlaxAlbertModel.from_pretrained("albert-base-v2" ) a : Union[str, Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) a : int = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) a : Any = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] a : List[Any] = (1, 11, 768) self.assertEqual(output.shape , lowerCAmelCase__ ) a : Optional[int] = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCAmelCase__ , atol=1E-4 ) )
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Union[str, Any] = { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/config.json''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/config.json''', } class __UpperCamelCase ( a__ ): lowerCamelCase : str ="""xlnet""" lowerCamelCase : Dict =["""mems"""] lowerCamelCase : Optional[int] ={ """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase__=3_2000 , lowerCAmelCase__=1024 , lowerCAmelCase__=24 , lowerCAmelCase__=16 , lowerCAmelCase__=4096 , lowerCAmelCase__="gelu" , lowerCAmelCase__=True , lowerCAmelCase__="bi" , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=-1 , lowerCAmelCase__=False , lowerCAmelCase__="last" , lowerCAmelCase__=True , lowerCAmelCase__="tanh" , lowerCAmelCase__=0.1 , lowerCAmelCase__=5 , lowerCAmelCase__=5 , lowerCAmelCase__=5 , lowerCAmelCase__=1 , lowerCAmelCase__=2 , **lowerCAmelCase__ , ) -> Optional[int]: a : Any = vocab_size a : Tuple = d_model a : Tuple = n_layer a : Tuple = n_head if d_model % n_head != 0: raise ValueError(f"""'d_model % n_head' ({d_model % n_head}) should be equal to 0""" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( f"""`d_head` ({kwargs['d_head']}) should be equal to `d_model // n_head` ({d_model // n_head})""" ) a : Tuple = d_model // n_head a : Optional[Any] = ff_activation a : List[Any] = d_inner a : Any = untie_r a : str = attn_type a : List[Any] = initializer_range a : Any = layer_norm_eps a : List[Any] = dropout a : List[Any] = mem_len a : Optional[Any] = reuse_len a : List[str] = bi_data a : Dict = clamp_len a : Any = same_length a : int = summary_type a : str = summary_use_proj a : Tuple = summary_activation a : Tuple = summary_last_dropout a : List[Any] = start_n_top a : Tuple = end_n_top a : Dict = bos_token_id a : List[Any] = pad_token_id a : Dict = eos_token_id if "use_cache" in kwargs: warnings.warn( "The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`" " instead." , lowerCAmelCase__ , ) a : Union[str, Any] = kwargs["use_cache"] a : Optional[Any] = use_mems_eval a : Dict = use_mems_train super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def __a ( self ) -> Union[str, Any]: logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def __a ( self , lowerCAmelCase__ ) -> List[str]: # Message copied from Transformer-XL documentation raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __UpperCamelCase ( a__ ): lowerCamelCase : List[Any] =["""image_processor""", """tokenizer"""] lowerCamelCase : int ="""BlipImageProcessor""" lowerCamelCase : List[Any] =("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: a : int = False super().__init__(lowerCAmelCase__ , lowerCAmelCase__ ) a : Tuple = self.image_processor def __call__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 0 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = True , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> BatchEncoding: if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: a : int = self.tokenizer a : int = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) return text_encoding # add pixel_values a : Union[str, Any] = self.image_processor(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) if text is not None: a : List[Any] = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) else: a : List[str] = None if text_encoding is not None: encoding_image_processor.update(lowerCAmelCase__ ) return encoding_image_processor def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> str: return self.tokenizer.batch_decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: return self.tokenizer.decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) @property def __a ( self ) -> Any: a : Dict = self.tokenizer.model_input_names a : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging a : str = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple=None , _lowercase : Dict=None ) ->Dict: '''simple docstring''' return field(default_factory=lambda: default , metadata=_lowercase ) @dataclass class __UpperCamelCase : lowerCamelCase : List[str] =list_field( default=[] , metadata={ """help""": ( """Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version""" """ of all available models""" ) } , ) lowerCamelCase : List[int] =list_field( default=[8] , metadata={"""help""": """List of batch sizes for which memory and time performance will be evaluated"""} ) lowerCamelCase : List[int] =list_field( default=[8, 32, 128, 512] , metadata={"""help""": """List of sequence lengths for which memory and time performance will be evaluated"""} , ) lowerCamelCase : bool =field( default=a__ , metadata={"""help""": """Whether to benchmark inference of model. Inference can be disabled via --no-inference."""} , ) lowerCamelCase : bool =field( default=a__ , metadata={"""help""": """Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."""} , ) lowerCamelCase : bool =field( default=a__ , metadata={"""help""": """Whether to run on available tpu devices. TPU can be disabled via --no-tpu."""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Use FP16 to accelerate inference."""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Benchmark training of model"""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Verbose memory tracing"""} ) lowerCamelCase : bool =field( default=a__ , metadata={"""help""": """Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."""} , ) lowerCamelCase : bool =field( default=a__ , metadata={ """help""": """Whether to perform memory measurements. Memory measurements can be disabled via --no-memory""" } , ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Trace memory line by line"""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Save result to a CSV file"""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Save all print statements in a log file"""} ) lowerCamelCase : bool =field(default=a__ , metadata={"""help""": """Whether to print environment information"""} ) lowerCamelCase : bool =field( default=a__ , metadata={ """help""": ( """Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use""" """ multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled""" """ for debugging / testing and on TPU.""" ) } , ) lowerCamelCase : str =field( default=F"""inference_time_{round(time() )}.csv""" , metadata={"""help""": """CSV filename used if saving time results to csv."""} , ) lowerCamelCase : str =field( default=F"""inference_memory_{round(time() )}.csv""" , metadata={"""help""": """CSV filename used if saving memory results to csv."""} , ) lowerCamelCase : str =field( default=F"""train_time_{round(time() )}.csv""" , metadata={"""help""": """CSV filename used if saving time results to csv for training."""} , ) lowerCamelCase : str =field( default=F"""train_memory_{round(time() )}.csv""" , metadata={"""help""": """CSV filename used if saving memory results to csv for training."""} , ) lowerCamelCase : str =field( default=F"""env_info_{round(time() )}.csv""" , metadata={"""help""": """CSV filename used if saving environment information."""} , ) lowerCamelCase : str =field( default=F"""log_{round(time() )}.csv""" , metadata={"""help""": """Log filename used if print statements are saved in log."""} , ) lowerCamelCase : int =field(default=3 , metadata={"""help""": """Times an experiment will be run."""} ) lowerCamelCase : bool =field( default=a__ , metadata={ """help""": ( """Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain""" """ model weights.""" ) } , ) def __a ( self ) -> Dict: warnings.warn( f"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" " are deprecated in general and it is advised to use external Benchmarking libraries " " to benchmark Transformer models." , lowerCAmelCase__ , ) def __a ( self ) -> Union[str, Any]: return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def __a ( self ) -> List[str]: if len(self.models ) <= 0: raise ValueError( "Please make sure you provide at least one model name / model identifier, *e.g.* `--models" " bert-base-cased` or `args.models = ['bert-base-cased']." ) return self.models @property def __a ( self ) -> List[Any]: if not self.multi_process: return False elif self.is_tpu: logger.info("Multiprocessing is currently not possible on TPU." ) return False else: return True
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=12 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=32 , lowerCAmelCase__=2 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=0.02 , lowerCAmelCase__=0 , lowerCAmelCase__=None , ) -> List[Any]: a : List[str] = parent a : Dict = batch_size a : List[Any] = seq_length a : int = is_training a : str = use_input_mask a : str = use_labels a : Dict = vocab_size a : Tuple = hidden_size a : Tuple = projection_dim a : List[str] = num_hidden_layers a : Dict = num_attention_heads a : str = intermediate_size a : str = dropout a : Union[str, Any] = attention_dropout a : Optional[int] = max_position_embeddings a : Union[str, Any] = initializer_range a : Dict = scope a : Union[str, Any] = bos_token_id def __a ( self ) -> Dict: a : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : Any = None if self.use_input_mask: a : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: a : Dict = input_mask.numpy() a, a : Union[str, Any] = input_mask.shape a : str = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): a : Optional[int] = 1 a : Union[str, Any] = 0 a : int = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def __a ( self ) -> str: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) a : Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , training=lowerCAmelCase__ ) a : List[str] = model(lowerCAmelCase__ , training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __a ( self ) -> Optional[Any]: a : Optional[Any] = self.prepare_config_and_inputs() a, a, a : Tuple = config_and_inputs a : Any = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : int =(TFBlipTextModel,) if is_tf_available() else () lowerCamelCase : Any =False lowerCamelCase : Union[str, Any] =False lowerCamelCase : Any =False def __a ( self ) -> Optional[Any]: a : Any = BlipTextModelTester(self ) a : Dict = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=37 ) def __a ( self ) -> Optional[int]: self.config_tester.run_common_tests() def __a ( self ) -> List[str]: a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __a ( self ) -> Dict: pass def __a ( self ) -> Tuple: pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def __a ( self ) -> Optional[Any]: pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def __a ( self ) -> List[Any]: pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def __a ( self ) -> List[Any]: pass @slow def __a ( self ) -> Optional[int]: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__=True ) -> Any: super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Tuple = logging.get_logger(__name__) a : List[str] = '''▁''' a : Optional[int] = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', } a : Optional[Any] = { '''vocab_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/vocab.json''' ), }, '''spm_file''': { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/sentencepiece.bpe.model''' ) }, } a : Tuple = { '''facebook/s2t-small-librispeech-asr''': 1024, } a : Optional[int] = ['''pt''', '''fr''', '''ru''', '''nl''', '''ro''', '''it''', '''es''', '''de'''] a : str = {'''mustc''': MUSTC_LANGS} class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : List[Any] =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =MAX_MODEL_INPUT_SIZES lowerCamelCase : List[str] =["""input_ids""", """attention_mask"""] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> None: a : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , do_upper_case=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , lang_codes=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase__ , ) a : Tuple = do_upper_case a : List[Any] = do_lower_case a : List[str] = load_json(lowerCAmelCase__ ) a : str = {v: k for k, v in self.encoder.items()} a : Tuple = spm_file a : int = load_spm(lowerCAmelCase__ , self.sp_model_kwargs ) if lang_codes is not None: a : List[Any] = lang_codes a : Any = LANGUAGES[lang_codes] a : Tuple = [f"""<lang:{lang}>""" for lang in self.langs] a : Tuple = {lang: self.sp_model.PieceToId(f"""<lang:{lang}>""" ) for lang in self.langs} a : Dict = self.lang_tokens a : Union[str, Any] = tgt_lang if tgt_lang is not None else self.langs[0] self.set_tgt_lang_special_tokens(self._tgt_lang ) else: a : Tuple = {} @property def __a ( self ) -> int: return len(self.encoder ) @property def __a ( self ) -> str: return self._tgt_lang @tgt_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_tgt_lang self.set_tgt_lang_special_tokens(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.lang_code_to_id[tgt_lang] a : Union[str, Any] = [lang_code_id] def __a ( self , lowerCAmelCase__ ) -> List[str]: return self.sp_model.encode(lowerCAmelCase__ , out_type=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> Any: return self.encoder.get(lowerCAmelCase__ , self.encoder[self.unk_token] ) def __a ( self , lowerCAmelCase__ ) -> str: return self.decoder.get(lowerCAmelCase__ , self.unk_token ) def __a ( self , lowerCAmelCase__ ) -> str: a : int = [] a : Optional[int] = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: a : Optional[Any] = self.sp_model.decode(lowerCAmelCase__ ) out_string += (decoded.upper() if self.do_upper_case else decoded) + token + " " a : Any = [] else: current_sub_tokens.append(lowerCAmelCase__ ) a : Optional[int] = self.sp_model.decode(lowerCAmelCase__ ) out_string += decoded.upper() if self.do_upper_case else decoded return out_string.strip() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + [self.eos_token_id] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) a : int = [1] * len(self.prefix_tokens ) a : Tuple = [1] if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__ )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__ )) + ([0] * len(lowerCAmelCase__ )) + suffix_ones def __a ( self ) -> Dict: a : Optional[Any] = self.encoder.copy() vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Dict: a : Optional[Any] = self.__dict__.copy() a : Dict = None return state def __setstate__( self , lowerCAmelCase__ ) -> None: a : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): a : Optional[int] = {} a : Union[str, Any] = load_spm(self.spm_file , self.sp_model_kwargs ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: a : str = Path(lowerCAmelCase__ ) assert save_dir.is_dir(), f"""{save_directory} should be a directory""" a : Any = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) a : Tuple = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , lowerCAmelCase__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(lowerCAmelCase__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , lowerCAmelCase__ ) elif not os.path.isfile(self.spm_file ): with open(lowerCAmelCase__ , "wb" ) as fi: a : Optional[int] = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__ ) return (str(lowerCAmelCase__ ), str(lowerCAmelCase__ )) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : Dict[str, Any] ) ->sentencepiece.SentencePieceProcessor: '''simple docstring''' a : Optional[int] = sentencepiece.SentencePieceProcessor(**_lowercase ) spm.Load(str(_lowercase ) ) return spm def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->Union[Dict, List]: '''simple docstring''' with open(_lowercase , "r" ) as f: return json.load(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : str ) ->None: '''simple docstring''' with open(_lowercase , "w" ) as f: json.dump(_lowercase , _lowercase , indent=2 )
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" import string from math import logaa def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : str ) ->int: '''simple docstring''' a : Tuple = document.translate( str.maketrans("" , "" , string.punctuation ) ).replace("\n" , "" ) a : Union[str, Any] = document_without_punctuation.split(" " ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : str ) ->tuple[int, int]: '''simple docstring''' a : Any = corpus.lower().translate( str.maketrans("" , "" , string.punctuation ) ) # strip all punctuation and replace it with '' a : int = corpus_without_punctuation.split("\n" ) a : Dict = term.lower() return (len([doc for doc in docs if term in doc] ), len(_lowercase )) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : Union[str, Any]=False ) ->float: '''simple docstring''' if smoothing: if n == 0: raise ValueError("log10(0) is undefined." ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError("df must be > 0" ) elif n == 0: raise ValueError("log10(0) is undefined." ) return round(logaa(n / df ) , 3 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->float: '''simple docstring''' return round(tf * idf , 3 )
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 50 ) ->int: '''simple docstring''' a : Dict = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F'''{solution() = }''')
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" a : Tuple = { '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : int = { '''google/efficientnet-b7''': '''https://huggingface.co/google/efficientnet-b7/resolve/main/config.json''', } class __UpperCamelCase ( a__ ): lowerCamelCase : Any ="""efficientnet""" def __init__( self , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 600 , lowerCAmelCase__ = 2.0 , lowerCAmelCase__ = 3.1 , lowerCAmelCase__ = 8 , lowerCAmelCase__ = [3, 3, 5, 3, 5, 5, 3] , lowerCAmelCase__ = [32, 16, 24, 40, 80, 112, 192] , lowerCAmelCase__ = [16, 24, 40, 80, 112, 192, 320] , lowerCAmelCase__ = [] , lowerCAmelCase__ = [1, 2, 2, 2, 1, 2, 1] , lowerCAmelCase__ = [1, 2, 2, 3, 3, 4, 1] , lowerCAmelCase__ = [1, 6, 6, 6, 6, 6, 6] , lowerCAmelCase__ = 0.25 , lowerCAmelCase__ = "swish" , lowerCAmelCase__ = 2560 , lowerCAmelCase__ = "mean" , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 0.001 , lowerCAmelCase__ = 0.99 , lowerCAmelCase__ = 0.5 , lowerCAmelCase__ = 0.2 , **lowerCAmelCase__ , ) -> Dict: super().__init__(**lowerCAmelCase__ ) a : Optional[Any] = num_channels a : List[str] = image_size a : Any = width_coefficient a : Optional[int] = depth_coefficient a : Tuple = depth_divisor a : Any = kernel_sizes a : Any = in_channels a : int = out_channels a : List[Any] = depthwise_padding a : Optional[Any] = strides a : List[Any] = num_block_repeats a : List[Any] = expand_ratios a : List[str] = squeeze_expansion_ratio a : List[Any] = hidden_act a : List[Any] = hidden_dim a : int = pooling_type a : Any = initializer_range a : List[Any] = batch_norm_eps a : Dict = batch_norm_momentum a : int = dropout_rate a : List[str] = drop_connect_rate a : Dict = sum(lowerCAmelCase__ ) * 4 class __UpperCamelCase ( a__ ): lowerCamelCase : List[str] =version.parse("""1.11""" ) @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def __a ( self ) -> float: return 1E-5
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : Optional[int] = logging.get_logger(__name__) a : Any = {'''openai-gpt''': '''https://huggingface.co/openai-gpt/resolve/main/config.json'''} class __UpperCamelCase ( a__ ): lowerCamelCase : List[str] ="""openai-gpt""" lowerCamelCase : Tuple ={ """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase__=4_0478 , lowerCAmelCase__=512 , lowerCAmelCase__=768 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0.02 , lowerCAmelCase__="cls_index" , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=0.1 , **lowerCAmelCase__ , ) -> List[Any]: a : Any = vocab_size a : Union[str, Any] = n_positions a : Optional[Any] = n_embd a : Dict = n_layer a : Dict = n_head a : List[str] = afn a : str = resid_pdrop a : Any = embd_pdrop a : List[str] = attn_pdrop a : Dict = layer_norm_epsilon a : List[str] = initializer_range a : Union[str, Any] = summary_type a : Union[str, Any] = summary_use_proj a : Any = summary_activation a : Any = summary_first_dropout a : Union[str, Any] = summary_proj_to_labels super().__init__(**lowerCAmelCase__ )
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] ) ->float: '''simple docstring''' if not nums: # Makes sure that the list is not empty raise ValueError("List is empty" ) a : str = sum(_lowercase ) / len(_lowercase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(_lowercase ) if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] , _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): a, a : Any = array[indexa], array[indexa] def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] , _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' if length > 1: a : Tuple = int(length / 2 ) for i in range(_lowercase , low + middle ): comp_and_swap(_lowercase , _lowercase , i + middle , _lowercase ) bitonic_merge(_lowercase , _lowercase , _lowercase , _lowercase ) bitonic_merge(_lowercase , low + middle , _lowercase , _lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] , _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' if length > 1: a : Union[str, Any] = int(length / 2 ) bitonic_sort(_lowercase , _lowercase , _lowercase , 1 ) bitonic_sort(_lowercase , low + middle , _lowercase , 0 ) bitonic_merge(_lowercase , _lowercase , _lowercase , _lowercase ) if __name__ == "__main__": a : Tuple = input('''Enter numbers separated by a comma:\n''').strip() a : List[Any] = [int(item.strip()) for item in user_input.split(''',''')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('''\nSorted array in ascending order is: ''', end='''''') print(*unsorted, sep=''', ''') bitonic_merge(unsorted, 0, len(unsorted), 0) print('''Sorted array in descending order is: ''', end='''''') print(*unsorted, sep=''', ''')
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch("socket.socket" ) @patch("builtins.open" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[Any] ) ->Optional[Any]: '''simple docstring''' a : Optional[Any] = Mock() a : int = conn, Mock() a : Optional[int] = iter([1, None] ) a : Optional[int] = lambda _lowercase : next(_lowercase ) # ===== invoke ===== send_file(filename="mytext.txt" , testing=_lowercase ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class __UpperCamelCase ( a__ ): lowerCamelCase : List[str] ="""""" lowerCamelCase : Tuple ="""hf-legacy""" # "hf://"" is reserved for hffs def __init__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> str: super().__init__(self , **lowerCAmelCase__ ) a : int = repo_info a : Dict = token a : Union[str, Any] = None def __a ( self ) -> List[Any]: if self.dir_cache is None: a : Optional[Any] = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes a : List[str] = { "name": hf_file.rfilename, "size": None, "type": "file", } self.dir_cache.update( { str(lowerCAmelCase__ ): {"name": str(lowerCAmelCase__ ), "size": None, "type": "directory"} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "rb" , **lowerCAmelCase__ , ) -> List[str]: if not isinstance(self.repo_info , lowerCAmelCase__ ): raise NotImplementedError(f"""Open is only implemented for dataset repositories, but got {self.repo_info}""" ) a : Union[str, Any] = hf_hub_url(self.repo_info.id , lowerCAmelCase__ , revision=self.repo_info.sha ) return fsspec.open( lowerCAmelCase__ , mode=lowerCAmelCase__ , headers=get_authentication_headers_for_url(lowerCAmelCase__ , use_auth_token=self.token ) , client_kwargs={"trust_env": True} , ).open() def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[int]: self._get_dirs() a : Any = self._strip_protocol(lowerCAmelCase__ ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=False , **lowerCAmelCase__ ) -> List[str]: self._get_dirs() a : Optional[int] = PurePosixPath(path.strip("/" ) ) a : Dict = {} for p, f in self.dir_cache.items(): a : str = PurePosixPath(p.strip("/" ) ) a : Tuple = p.parent if root == path: a : Union[str, Any] = f a : List[Any] = list(paths.values() ) if detail: return out else: return sorted(f["name"] for f in out )
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" from __future__ import annotations from typing import Any class __UpperCamelCase : def __init__( self , lowerCAmelCase__ = 6 ) -> None: a : Node | None = None a : Node | None = None self.create_linked_list(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> None: a : Dict = Node() a : Union[str, Any] = current_node a : str = current_node a : str = current_node for _ in range(1 , lowerCAmelCase__ ): a : int = Node() a : Any = current_node a : Optional[int] = previous_node a : Dict = current_node a : Optional[Any] = self.front a : Tuple = previous_node def __a ( self ) -> bool: return ( self.front == self.rear and self.front is not None and self.front.data is None ) def __a ( self ) -> Any | None: self.check_can_perform_operation() return self.front.data if self.front else None def __a ( self , lowerCAmelCase__ ) -> None: if self.rear is None: return self.check_is_full() if not self.is_empty(): a : int = self.rear.next if self.rear: a : Union[str, Any] = data def __a ( self ) -> Any: self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: a : Tuple = self.front.data a : List[str] = None return data a : List[str] = self.front a : Any = old_front.next a : Optional[Any] = old_front.data a : int = None return data def __a ( self ) -> None: if self.is_empty(): raise Exception("Empty Queue" ) def __a ( self ) -> None: if self.rear and self.rear.next == self.front: raise Exception("Full Queue" ) class __UpperCamelCase : def __init__( self ) -> None: a : Any | None = None a : Node | None = None a : Node | None = None if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=sys.maxsize ) -> str: a : List[str] = "bilinear" a : str = max_size a : Union[str, Any] = short_edge_length def __call__( self , lowerCAmelCase__ ) -> Union[str, Any]: a : List[Any] = [] for img in imgs: a, a : int = img.shape[:2] # later: provide list and randomly choose index for resize a : str = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img a : Dict = size * 1.0 / min(lowerCAmelCase__ , lowerCAmelCase__ ) if h < w: a, a : Dict = size, scale * w else: a, a : List[Any] = scale * h, size if max(lowerCAmelCase__ , lowerCAmelCase__ ) > self.max_size: a : int = self.max_size * 1.0 / max(lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = newh * scale a : Optional[int] = neww * scale a : int = int(neww + 0.5 ) a : int = int(newh + 0.5 ) if img.dtype == np.uinta: a : Optional[int] = Image.fromarray(lowerCAmelCase__ ) a : Any = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) a : Optional[Any] = np.asarray(lowerCAmelCase__ ) else: a : Dict = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw a : Optional[Any] = nn.functional.interpolate( lowerCAmelCase__ , (newh, neww) , mode=self.interp_method , align_corners=lowerCAmelCase__ ).squeeze(0 ) img_augs.append(lowerCAmelCase__ ) return img_augs class __UpperCamelCase : def __init__( self , lowerCAmelCase__ ) -> Optional[Any]: a : Any = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) a : Any = cfg.INPUT.FORMAT a : Optional[Any] = cfg.SIZE_DIVISIBILITY a : List[str] = cfg.PAD_VALUE a : List[str] = cfg.INPUT.MAX_SIZE_TEST a : Optional[int] = cfg.MODEL.DEVICE a : List[Any] = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) a : Union[str, Any] = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) a : Dict = lambda lowerCAmelCase__ : (x - self.pixel_mean) / self.pixel_std def __a ( self , lowerCAmelCase__ ) -> str: a : int = tuple(max(lowerCAmelCase__ ) for s in zip(*[img.shape for img in images] ) ) a : Dict = [im.shape[-2:] for im in images] a : Optional[Any] = [ nn.functional.pad( lowerCAmelCase__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(lowerCAmelCase__ , lowerCAmelCase__ ) ] return torch.stack(lowerCAmelCase__ ), torch.tensor(lowerCAmelCase__ ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__=False ) -> List[Any]: with torch.no_grad(): if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Optional[Any] = [images] if single_image: assert len(lowerCAmelCase__ ) == 1 for i in range(len(lowerCAmelCase__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(lowerCAmelCase__ , images.pop(lowerCAmelCase__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( lowerCAmelCase__ , torch.as_tensor(img_tensorize(images.pop(lowerCAmelCase__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge a : int = torch.tensor([im.shape[:2] for im in images] ) a : List[str] = self.aug(lowerCAmelCase__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic a : Optional[int] = [self.normalizer(lowerCAmelCase__ ) for x in images] # now pad them to do the following operations a, a : List[Any] = self.pad(lowerCAmelCase__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad a : Union[str, Any] = torch.true_divide(lowerCAmelCase__ , lowerCAmelCase__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Any ) ->str: '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Tuple[int, int] ) ->Union[str, Any]: '''simple docstring''' assert torch.isfinite(_lowercase ).all(), "Box tensor contains infinite or NaN!" a, a : Tuple = box_size tensor[:, 0].clamp_(min=0 , max=_lowercase ) tensor[:, 1].clamp_(min=0 , max=_lowercase ) tensor[:, 2].clamp_(min=0 , max=_lowercase ) tensor[:, 3].clamp_(min=0 , max=_lowercase )
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" import argparse import os import re a : Any = '''src/transformers/models/auto''' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict a : Dict = re.compile(R'''[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict''') # re pattern that matches identifiers in mappings a : str = re.compile(R'''\s*\(\s*"(\S[^"]+)"''') def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : bool = False ) ->Optional[int]: '''simple docstring''' with open(_lowercase , "r" , encoding="utf-8" ) as f: a : Optional[int] = f.read() a : Dict = content.split("\n" ) a : List[str] = [] a : Tuple = 0 while line_idx < len(_lowercase ): if _re_intro_mapping.search(lines[line_idx] ) is not None: a : str = len(re.search(R"^(\s*)\S" , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(" " * indent + "(" ): new_lines.append(lines[line_idx] ) line_idx += 1 a : Optional[int] = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": a : int = line_idx while not lines[line_idx].startswith(" " * indent + ")" ): line_idx += 1 blocks.append("\n".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers a : List[str] = sorted(_lowercase , key=lambda _lowercase : _re_identifier.search(_lowercase ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(_lowercase , "w" , encoding="utf-8" ) as f: f.write("\n".join(_lowercase ) ) elif "\n".join(_lowercase ) != content: return True def _SCREAMING_SNAKE_CASE ( _lowercase : bool = False ) ->str: '''simple docstring''' a : List[str] = [os.path.join(_lowercase , _lowercase ) for f in os.listdir(_lowercase ) if f.endswith(".py" )] a : Any = [sort_auto_mapping(_lowercase , overwrite=_lowercase ) for fname in fnames] if not overwrite and any(_lowercase ): a : int = [f for f, d in zip(_lowercase , _lowercase ) if d] raise ValueError( F"""The following files have auto mappings that need sorting: {', '.join(_lowercase )}. Run `make style` to fix""" " this." ) if __name__ == "__main__": a : int = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') a : Optional[int] = parser.parse_args() sort_all_auto_mappings(not args.check_only)
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __a ( self ) -> Optional[Any]: a : str = 1 a : Tuple = 3 a : Tuple = (32, 32) a : Tuple = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def __a ( self ) -> Any: torch.manual_seed(0 ) a : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def __a ( self ) -> Union[str, Any]: torch.manual_seed(0 ) a : Dict = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def __a ( self ) -> str: torch.manual_seed(0 ) a : Union[str, Any] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def __a ( self ) -> Any: def extract(*lowerCAmelCase__ , **lowerCAmelCase__ ): class __UpperCamelCase : def __init__( self ) -> Optional[Any]: a : Dict = torch.ones([0] ) def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def __a ( self ) -> List[Any]: a : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator a : Any = self.dummy_cond_unet a : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) a : Tuple = self.dummy_vae a : Any = self.dummy_text_encoder a : Tuple = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) a : str = 77 a : Tuple = self.dummy_image.to(lowerCAmelCase__ ) a : Optional[int] = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk a : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) a : Any = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) a : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) a : List[Any] = "A painting of a squirrel eating a burger" a : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) a : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) a : List[str] = output.images a : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) a : List[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] a : Optional[int] = image[0, -3:, -3:, -1] a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) a : Dict = np.array([0.4_427, 0.3_731, 0.4_249, 0.4_941, 0.4_546, 0.4_148, 0.4_193, 0.4_666, 0.4_499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def __a ( self ) -> List[str]: a : str = self.dummy_cond_unet a : int = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) a : Optional[Any] = self.dummy_vae a : Tuple = self.dummy_text_encoder a : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) a : Optional[Any] = 77 a : Any = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 a : Any = unet.half() a : List[str] = vae.half() a : Optional[Any] = bert.half() # make sure here that pndm scheduler skips prk a : List[str] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) a : Any = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) a : List[str] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) a : Any = "A painting of a squirrel eating a burger" a : Any = torch.manual_seed(0 ) a : int = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def __a ( self ) -> List[Any]: a : List[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 a : Optional[Any] = init_image.resize((760, 504) ) a : Optional[int] = "BAAI/AltDiffusion" a : Union[str, Any] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() a : str = "A fantasy landscape, trending on artstation" a : Tuple = torch.manual_seed(0 ) a : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.75 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) a : Union[str, Any] = output.images[0] a : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) a : Dict = np.array([0.9_358, 0.9_397, 0.9_599, 0.9_901, 1.0_000, 1.0_000, 0.9_882, 1.0_000, 1.0_000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Optional[int]: a : str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) a : Any = init_image.resize((768, 512) ) a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) a : Optional[Any] = "BAAI/AltDiffusion" a : Optional[int] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() a : Tuple = "A fantasy landscape, trending on artstation" a : Union[str, Any] = torch.manual_seed(0 ) a : Dict = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.75 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) a : int = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" from scipy.stats import pearsonr import datasets a : Union[str, Any] = ''' Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. ''' a : Union[str, Any] = ''' Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results[\'pearsonr\'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) [\'p-value\', \'pearsonr\'] >>> print(round(results[\'pearsonr\'], 2)) -0.74 >>> print(round(results[\'p-value\'], 2)) 0.15 ''' a : List[Any] = ''' @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def __a ( self ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"] , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ) -> List[Any]: if return_pvalue: a : Tuple = pearsonr(lowerCAmelCase__ , lowerCAmelCase__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCAmelCase__ , lowerCAmelCase__ )[0] )}
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor a : Tuple = logging.get_logger(__name__) class __UpperCamelCase ( a__ ): def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> None: warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __UpperCamelCase ( a__ ): def __a ( self ) -> Optional[Any]: a : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "num_attention_heads" ) ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=64 , lowerCAmelCase__=3 , lowerCAmelCase__=3 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=16 , lowerCAmelCase__=[128, 256, 384] , lowerCAmelCase__=[4, 6, 8] , lowerCAmelCase__=[2, 3, 4] , lowerCAmelCase__=[16, 16, 16] , lowerCAmelCase__=0 , lowerCAmelCase__=[2, 2, 2] , lowerCAmelCase__=[2, 2, 2] , lowerCAmelCase__=0.02 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=2 , ) -> Union[str, Any]: a : List[str] = parent a : List[str] = batch_size a : Union[str, Any] = image_size a : str = num_channels a : Optional[Any] = kernel_size a : Optional[int] = stride a : List[Any] = padding a : Tuple = hidden_sizes a : Any = num_attention_heads a : Union[str, Any] = depths a : int = key_dim a : Union[str, Any] = drop_path_rate a : Optional[Any] = patch_size a : int = attention_ratio a : str = mlp_ratio a : Any = initializer_range a : Union[str, Any] = [ ["Subsample", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["Subsample", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] a : Dict = is_training a : Tuple = use_labels a : Any = num_labels a : Any = initializer_range def __a ( self ) -> Tuple: a : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a : Tuple = None if self.use_labels: a : List[str] = ids_tensor([self.batch_size] , self.num_labels ) a : Union[str, Any] = self.get_config() return config, pixel_values, labels def __a ( self ) -> Tuple: return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: a : Union[str, Any] = LevitModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : List[str] = model(lowerCAmelCase__ ) a : int = (self.image_size, self.image_size) a, a : Dict = image_size[0], image_size[1] for _ in range(4 ): a : int = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) a : Optional[int] = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: a : Optional[int] = self.num_labels a : str = LevitForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[Any] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self ) -> Tuple: a : List[str] = self.prepare_config_and_inputs() a, a, a : Optional[Any] = config_and_inputs a : Any = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Optional[Any] =( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) lowerCamelCase : Dict =( { """feature-extraction""": LevitModel, """image-classification""": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase : Union[str, Any] =False lowerCamelCase : int =False lowerCamelCase : Any =False lowerCamelCase : int =False lowerCamelCase : str =False def __a ( self ) -> Any: a : List[Any] = LevitModelTester(self ) a : Optional[int] = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=37 ) def __a ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __a ( self ) -> Tuple: return @unittest.skip(reason="Levit does not use inputs_embeds" ) def __a ( self ) -> Optional[Any]: pass @unittest.skip(reason="Levit does not support input and output embeddings" ) def __a ( self ) -> Dict: pass @unittest.skip(reason="Levit does not output attentions" ) def __a ( self ) -> Any: pass def __a ( self ) -> Dict: a, a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : List[Any] = model_class(lowerCAmelCase__ ) a : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : Optional[int] = [*signature.parameters.keys()] a : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def __a ( self ) -> int: def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): a : Optional[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): a : Optional[int] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Tuple = outputs.hidden_states a : Any = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) a : List[Any] = (self.model_tester.image_size, self.model_tester.image_size) a, a : Optional[int] = image_size[0], image_size[1] for _ in range(4 ): a : Optional[int] = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) a : str = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) a, a : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : List[str] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a : List[Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> List[Any]: pass def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ) -> Union[str, Any]: a : Optional[int] = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __a ( self ) -> Any: a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __a ( self ) -> str: a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: if not self.model_tester.is_training: return a, a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() a : List[Any] = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCAmelCase__ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue a : Any = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() a : int = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) a : Union[str, Any] = model(**lowerCAmelCase__ ).loss loss.backward() def __a ( self ) -> List[str]: a, a : str = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return a : Any = False a : str = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase__ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue a : Any = model_class(lowerCAmelCase__ ) model.gradient_checkpointing_enable() model.to(lowerCAmelCase__ ) model.train() a : str = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) a : Any = model(**lowerCAmelCase__ ).loss loss.backward() def __a ( self ) -> Dict: a, a : str = self.model_tester.prepare_config_and_inputs_for_common() a : str = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCAmelCase__ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): a : str = problem_type["title"] a : int = problem_type["num_labels"] a : Union[str, Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() a : List[Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) if problem_type["num_labels"] > 1: a : Optional[Any] = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) a : Optional[Any] = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCAmelCase__ ) as warning_list: a : Optional[int] = model(**lowerCAmelCase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def __a ( self ) -> Dict: for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : List[str] = LevitModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' a : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def __a ( self ) -> Union[str, Any]: return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __a ( self ) -> Union[str, Any]: a : Optional[int] = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowerCAmelCase__ ) a : Optional[Any] = self.default_image_processor a : int = prepare_img() a : Any = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : Optional[Any] = model(**lowerCAmelCase__ ) # verify the logits a : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) a : Optional[Any] = torch.tensor([1.0_448, -0.3_745, -1.8_317] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=2 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=10 , lowerCAmelCase__=3 , lowerCAmelCase__=32 * 8 , lowerCAmelCase__=32 * 8 , lowerCAmelCase__=4 , lowerCAmelCase__=64 , ) -> Tuple: a : List[str] = parent a : Optional[int] = batch_size a : str = is_training a : List[str] = use_auxiliary_loss a : Dict = num_queries a : int = num_channels a : int = min_size a : Optional[Any] = max_size a : str = num_labels a : List[str] = hidden_dim a : Optional[int] = hidden_dim def __a ( self ) -> Union[str, Any]: a : List[str] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowerCAmelCase__ ) a : List[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] , device=lowerCAmelCase__ ) a : Any = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=lowerCAmelCase__ ) > 0.5 ).float() a : Optional[Any] = (torch.rand((self.batch_size, self.num_labels) , device=lowerCAmelCase__ ) > 0.5).long() a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __a ( self ) -> Tuple: a : str = MaskaFormerConfig( hidden_size=self.hidden_dim , ) a : Any = self.num_queries a : str = self.num_labels a : Union[str, Any] = [1, 1, 1, 1] a : str = self.num_channels a : List[Any] = 64 a : Any = 128 a : Optional[int] = self.hidden_dim a : List[Any] = self.hidden_dim a : Optional[Any] = self.hidden_dim return config def __a ( self ) -> Optional[int]: a, a, a, a, a : List[str] = self.prepare_config_and_inputs() a : int = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : Tuple = output.encoder_hidden_states a : Optional[int] = output.pixel_decoder_hidden_states a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , config.decoder_layers ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ) -> Optional[Any]: with torch.no_grad(): a : List[Any] = MaskaFormerModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[int] = model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) a : Union[str, Any] = model(lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: a : Any = MaskaFormerForUniversalSegmentation(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() def comm_check_on_output(lowerCAmelCase__ ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): a : List[str] = model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) a : Union[str, Any] = model(lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) a : Tuple = model( pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Optional[Any] =(MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () lowerCamelCase : List[str] ={"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} lowerCamelCase : int =False lowerCamelCase : Dict =False lowerCamelCase : Any =False lowerCamelCase : List[Any] =False def __a ( self ) -> int: a : str = MaskaFormerModelTester(self ) a : Optional[Any] = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def __a ( self ) -> int: a, a : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def __a ( self ) -> Any: a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*lowerCAmelCase__ ) @unittest.skip(reason="Mask2Former does not use inputs_embeds" ) def __a ( self ) -> Dict: pass @unittest.skip(reason="Mask2Former does not have a get_input_embeddings method" ) def __a ( self ) -> Any: pass @unittest.skip(reason="Mask2Former is not a generative model" ) def __a ( self ) -> Any: pass @unittest.skip(reason="Mask2Former does not use token embeddings" ) def __a ( self ) -> List[Any]: pass @require_torch_multi_gpu @unittest.skip( reason="Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def __a ( self ) -> List[str]: pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> int: pass def __a ( self ) -> int: a, a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Optional[Any] = model_class(lowerCAmelCase__ ) a : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : Dict = [*signature.parameters.keys()] a : Tuple = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) @slow def __a ( self ) -> List[Any]: for model_name in ["facebook/mask2former-swin-small-coco-instance"]: a : Optional[Any] = MaskaFormerModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Union[str, Any] = (self.model_tester.min_size,) * 2 a : List[str] = { "pixel_values": torch.randn((2, 3, *size) , device=lowerCAmelCase__ ), "mask_labels": torch.randn((2, 10, *size) , device=lowerCAmelCase__ ), "class_labels": torch.zeros(2 , 10 , device=lowerCAmelCase__ ).long(), } a : List[Any] = self.model_tester.get_config() a : Union[str, Any] = MaskaFormerForUniversalSegmentation(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[Any] = model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None ) def __a ( self ) -> Optional[int]: a, a : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def __a ( self ) -> int: a, a : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Union[str, Any] = model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : Union[str, Any] = model(**lowerCAmelCase__ , output_attentions=lowerCAmelCase__ ) self.assertTrue(outputs.attentions is not None ) def __a ( self ) -> int: if not self.model_tester.is_training: return a : Union[str, Any] = self.all_model_classes[1] a, a, a, a, a : Optional[int] = self.model_tester.prepare_config_and_inputs() a : List[str] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() a : List[str] = model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ).loss loss.backward() def __a ( self ) -> List[str]: a : int = self.all_model_classes[1] a, a, a, a, a : int = self.model_tester.prepare_config_and_inputs() a : Tuple = True a : int = True a : Optional[int] = model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) model.train() a : int = model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) a : List[Any] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() a : List[str] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() a : Union[str, Any] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() a : Tuple = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowerCAmelCase__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) a : Tuple = 1E-4 def _SCREAMING_SNAKE_CASE ( ) ->str: '''simple docstring''' a : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class __UpperCamelCase ( unittest.TestCase ): @cached_property def __a ( self ) -> Optional[int]: return "facebook/mask2former-swin-small-coco-instance" @cached_property def __a ( self ) -> Dict: return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __a ( self ) -> str: a : str = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ) a : int = self.default_image_processor a : Tuple = prepare_img() a : Optional[int] = image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) a : Any = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 384, 384) ) with torch.no_grad(): a : int = model(**lowerCAmelCase__ ) a : Union[str, Any] = torch.tensor( [[-0.2_790, -1.0_717, -1.1_668], [-0.5_128, -0.3_128, -0.4_987], [-0.5_832, 0.1_971, -0.0_197]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) a : Optional[int] = torch.tensor( [[0.8_973, 1.1_847, 1.1_776], [1.1_934, 1.5_040, 1.5_128], [1.1_153, 1.4_486, 1.4_951]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) a : Dict = torch.tensor( [[2.1_152, 1.7_000, -0.8_603], [1.5_808, 1.8_004, -0.9_353], [1.6_043, 1.7_495, -0.5_999]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def __a ( self ) -> int: a : Tuple = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ).eval() a : List[str] = self.default_image_processor a : Dict = prepare_img() a : Optional[int] = image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) a : int = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 384, 384) ) with torch.no_grad(): a : Optional[Any] = model(**lowerCAmelCase__ ) # masks_queries_logits a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) a : Any = [ [-8.7_839, -9.0_056, -8.8_121], [-7.4_104, -7.0_313, -6.5_401], [-6.6_105, -6.3_427, -6.4_675], ] a : Any = torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) # class_queries_logits a : Tuple = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) a : Union[str, Any] = torch.tensor( [ [1.8_324, -8.0_835, -4.1_922], [0.8_450, -9.0_050, -3.6_053], [0.3_045, -7.7_293, -3.0_275], ] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Optional[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase__ ).eval() a : List[Any] = self.default_image_processor a : Any = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) a : int = inputs["pixel_values"].to(lowerCAmelCase__ ) a : Union[str, Any] = [el.to(lowerCAmelCase__ ) for el in inputs["mask_labels"]] a : str = [el.to(lowerCAmelCase__ ) for el in inputs["class_labels"]] with torch.no_grad(): a : Dict = model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None )
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : list[list[int | float]] ) ->int: '''simple docstring''' a : List[Any] = len(_lowercase ) a : Dict = len(matrix[0] ) a : List[str] = min(_lowercase , _lowercase ) for row in range(_lowercase ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _lowercase ): a : str = matrix[col][row] / matrix[row][row] for i in range(_lowercase , _lowercase ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows a : List[Any] = True for i in range(row + 1 , _lowercase ): if matrix[i][row] != 0: a, a : Union[str, Any] = matrix[i], matrix[row] a : Union[str, Any] = False break if reduce: rank -= 1 for i in range(_lowercase ): a : Optional[int] = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class __UpperCamelCase : @property def __a ( self ) -> List[str]: return self.get_dummy_input() @property def __a ( self ) -> List[str]: if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f"""'{self.block_type}' is not a supported block_type. Set it to 'up', 'mid', or 'down'.""" ) def __a ( self , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , ) -> List[Any]: a : Any = 4 a : List[str] = 32 a : Dict = (32, 32) a : Union[str, Any] = torch.manual_seed(0 ) a : Dict = torch.device(lowerCAmelCase__ ) a : Tuple = (batch_size, num_channels) + sizes a : int = randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=lowerCAmelCase__ ) a : str = {"hidden_states": hidden_states} if include_temb: a : Dict = 128 a : Optional[int] = randn_tensor((batch_size, temb_channels) , generator=lowerCAmelCase__ , device=lowerCAmelCase__ ) if include_res_hidden_states_tuple: a : Optional[Any] = torch.manual_seed(1 ) a : Optional[int] = (randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=lowerCAmelCase__ ),) if include_encoder_hidden_states: a : List[str] = floats_tensor((batch_size, 32, 32) ).to(lowerCAmelCase__ ) if include_skip_sample: a : Tuple = randn_tensor(((batch_size, 3) + sizes) , generator=lowerCAmelCase__ , device=lowerCAmelCase__ ) return dummy_input def __a ( self ) -> Dict: a : List[Any] = { "in_channels": 32, "out_channels": 32, "temb_channels": 128, } if self.block_type == "up": a : Optional[Any] = 32 if self.block_type == "mid": init_dict.pop("out_channels" ) a : List[str] = self.dummy_input return init_dict, inputs_dict def __a ( self , lowerCAmelCase__ ) -> str: a, a : Tuple = self.prepare_init_args_and_inputs_for_common() a : Tuple = self.block_class(**lowerCAmelCase__ ) unet_block.to(lowerCAmelCase__ ) unet_block.eval() with torch.no_grad(): a : int = unet_block(**lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : str = output[0] self.assertEqual(output.shape , self.output_shape ) a : List[str] = output[0, -1, -3:, -3:] a : Union[str, Any] = torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) assert torch_all_close(output_slice.flatten() , lowerCAmelCase__ , atol=5E-3 ) @unittest.skipIf(torch_device == "mps" , "Training is not supported in mps" ) def __a ( self ) -> List[str]: a, a : Optional[Any] = self.prepare_init_args_and_inputs_for_common() a : Tuple = self.block_class(**lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() a : Optional[int] = model(**lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[str] = output[0] a : List[Any] = torch.device(lowerCAmelCase__ ) a : Any = randn_tensor(output.shape , device=lowerCAmelCase__ ) a : int = torch.nn.functional.mse_loss(lowerCAmelCase__ , lowerCAmelCase__ ) loss.backward()
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging a : str = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ = 101 ) -> List[Any]: a : Union[str, Any] = length def __len__( self ) -> List[str]: return self.length def __getitem__( self , lowerCAmelCase__ ) -> int: return i class __UpperCamelCase : def __call__( self , lowerCAmelCase__ ) -> int: return {"input_ids": torch.tensor(lowerCAmelCase__ ), "labels": torch.tensor(lowerCAmelCase__ )} class __UpperCamelCase ( nn.Module ): def __init__( self ) -> int: super().__init__() # Add some (unused) params otherwise DDP will complain. a : Any = nn.Linear(120 , 80 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> List[str]: if labels is not None: return torch.tensor(0.0 , device=input_ids.device ), input_ids else: return input_ids class __UpperCamelCase ( a__ ): @require_torch_neuroncore def __a ( self ) -> Union[str, Any]: a : Any = f"""--nproc_per_node=2 --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py """.split() a : str = self.get_auto_remove_tmp_dir() a : Union[str, Any] = f"""--output_dir {output_dir}""".split() a : Optional[int] = ["torchrun"] + distributed_args + args execute_subprocess_async(lowerCAmelCase__ , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call class __UpperCamelCase ( a__ ): @require_torch_multi_gpu def __a ( self ) -> List[str]: a : Any = f"""--nproc_per_node={torch.cuda.device_count()} --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py """.split() a : int = self.get_auto_remove_tmp_dir() a : int = f"""--output_dir {output_dir}""".split() a : Any = ["torchrun"] + distributed_args + args execute_subprocess_async(lowerCAmelCase__ , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py a : List[str] = HfArgumentParser((TrainingArguments,)) a : Tuple = parser.parse_args_into_dataclasses()[0] logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, ''' F'''distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}''' ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: a : List[Any] = DummyDataset(dataset_length) def _SCREAMING_SNAKE_CASE ( _lowercase : EvalPrediction ) ->Dict: '''simple docstring''' a : List[str] = list(range(len(_lowercase ) ) ) a : Union[str, Any] = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( "Predictions and/or labels do not match expected results:\n - predictions: " F"""{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}""" ) return {"success": success} a : List[str] = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) a : Tuple = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) a : int = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) a : Optional[Any] = 2 a : Optional[int] = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) a : Optional[int] = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) a : List[Any] = None
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Any = name a : Dict = value a : Union[str, Any] = weight def __repr__( self ) -> Optional[int]: return f"""{self.__class__.__name__}({self.name}, {self.value}, {self.weight})""" def __a ( self ) -> Any: return self.value def __a ( self ) -> int: return self.name def __a ( self ) -> Optional[Any]: return self.weight def __a ( self ) -> int: return self.value / self.weight def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : int , _lowercase : List[Any] ) ->Optional[int]: '''simple docstring''' a : List[str] = [] for i in range(len(_lowercase ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Optional[int] , _lowercase : str ) ->List[str]: '''simple docstring''' a : Dict = sorted(_lowercase , key=_lowercase , reverse=_lowercase ) a : Union[str, Any] = [] a, a : Dict = 0.0, 0.0 for i in range(len(_lowercase ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def _SCREAMING_SNAKE_CASE ( _lowercase : Callable[[int | float], int | float] , _lowercase : int | float , _lowercase : int | float , _lowercase : int = 100 , ) ->float: '''simple docstring''' a : str = x_start a : Optional[Any] = fnc(_lowercase ) a : Union[str, Any] = 0.0 for _ in range(_lowercase ): # Approximates curve as a sequence of linear lines and sums their length a : Optional[Any] = (x_end - x_start) / steps + xa a : str = fnc(_lowercase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step a : Tuple = xa a : Optional[Any] = fxa return length if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Optional[Any]: '''simple docstring''' return math.sin(10 * x ) print('''f(x) = sin(10 * x)''') print('''The length of the curve from x = -10 to x = 10 is:''') a : Dict = 10 while i <= 100000: print(F'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin a : str = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right a : List[Any] = 50003 a : Union[str, Any] = 50002 @require_sentencepiece @require_tokenizers class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : List[Any] =PLBartTokenizer lowerCamelCase : List[Any] =None lowerCamelCase : Union[str, Any] =False def __a ( self ) -> Union[str, Any]: super().setUp() # We have a SentencePiece fixture for testing a : List[str] = PLBartTokenizer(lowerCAmelCase__ , language_codes="base" , keep_accents=lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self ) -> Dict: a : Optional[int] = PLBartTokenizer(lowerCAmelCase__ , language_codes="base" , keep_accents=lowerCAmelCase__ ) a : Optional[int] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) a : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) a : Union[str, Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) a : Tuple = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) a : Optional[int] = tokenizer.vocab_size a : Any = [tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) for x in range(end - 4 , lowerCAmelCase__ )] self.assertListEqual(lowerCAmelCase__ , ["__java__", "__python__", "__en_XX__", "<mask>"] ) a : Any = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" a : Optional[int] = tokenizer(lowerCAmelCase__ ).input_ids self.assertEqual( tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ ) , lowerCAmelCase__ , ) def __a ( self ) -> Tuple: a : Any = PLBartTokenizer(lowerCAmelCase__ , language_codes="multi" , keep_accents=lowerCAmelCase__ ) a : Dict = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) a : int = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) a : str = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) a : Union[str, Any] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) a : Dict = tokenizer.vocab_size a : List[Any] = [tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) for x in range(end - 7 , lowerCAmelCase__ )] self.assertListEqual( lowerCAmelCase__ , ["__java__", "__python__", "__en_XX__", "__javascript__", "__php__", "__ruby__", "__go__"] ) a : List[str] = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" a : int = tokenizer(lowerCAmelCase__ ).input_ids self.assertEqual( tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ , clean_up_tokenization_spaces=lowerCAmelCase__ ) , lowerCAmelCase__ , ) @require_torch @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): lowerCamelCase : int ="""uclanlp/plbart-python-en_XX""" lowerCamelCase : Tuple =[ """def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])""", """def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])""", ] lowerCamelCase : List[Any] =[ """Returns the maximum value of a b c.""", """Sums the values of a b c.""", ] lowerCamelCase : List[Any] =[ 134, 5452, 3_3460, 3_3441, 3_3463, 3_3465, 3_3463, 3_3449, 988, 20, 3_3456, 19, 3_3456, 771, 39, 4258, 889, 3318, 3_3441, 3_3463, 3_3465, 3_3463, 3_3449, 2471, 2, PYTHON_CODE, ] @classmethod def __a ( cls ) -> List[Any]: a : PLBartTokenizer = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes="base" , src_lang="python" , tgt_lang="en_XX" ) a : Optional[Any] = 1 return cls def __a ( self ) -> List[str]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__java__"] , 5_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__python__"] , 5_0002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__en_XX__"] , 5_0003 ) def __a ( self ) -> int: a : List[str] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def __a ( self ) -> int: self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) a : List[str] = [EN_CODE, 9037, 3_3442, 57, 752, 153, 14, 56, 18, 9, 2] a : List[Any] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) a : Dict = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def __a ( self ) -> List[Any]: a : Union[str, Any] = ["def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])" * 20] self.assertIsInstance(src_text[0] , lowerCAmelCase__ ) a : Optional[Any] = 10 a : Tuple = self.tokenizer(lowerCAmelCase__ , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __a ( self ) -> Tuple: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "__java__"] ) , [5_0004, 5_0001] ) def __a ( self ) -> Tuple: a : str = tempfile.mkdtemp() a : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(lowerCAmelCase__ ) a : List[str] = PLBartTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , lowerCAmelCase__ ) @require_torch def __a ( self ) -> Optional[int]: a : List[Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) a : str = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , lowerCAmelCase__ ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def __a ( self ) -> Dict: a : Optional[int] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) a : Optional[int] = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual((2, 26) , batch.input_ids.shape ) self.assertEqual((2, 26) , batch.attention_mask.shape ) a : Union[str, Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def __a ( self ) -> Any: a : Optional[Any] = self.tokenizer(self.src_text , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=3 , return_tensors="pt" ) a : List[str] = self.tokenizer( text_target=self.tgt_text , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=10 , return_tensors="pt" ) a : List[Any] = targets["input_ids"] a : Tuple = shift_tokens_right(lowerCAmelCase__ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def __a ( self ) -> List[Any]: a : List[Any] = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="java" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # A, test, EOS, en_XX "input_ids": [[150, 242, 2, 5_0003]], "attention_mask": [[1, 1, 1, 1]], # java "forced_bos_token_id": 5_0001, } , )
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" class __UpperCamelCase : def __init__( self ) -> Any: a : Any = {} def __a ( self ) -> None: print(self.vertex ) for i in self.vertex: print(lowerCAmelCase__ , " -> " , " -> ".join([str(lowerCAmelCase__ ) for j in self.vertex[i]] ) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(lowerCAmelCase__ ) else: # else make a new vertex a : List[str] = [to_vertex] def __a ( self ) -> None: # visited array for storing already visited nodes a : List[str] = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: # mark start vertex as visited a : List[str] = True print(lowerCAmelCase__ , end=" " ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": a : Optional[Any] = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('''DFS:''') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" import random from .binary_exp_mod import bin_exp_mod def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Optional[int]=1000 ) ->str: '''simple docstring''' if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd a : Dict = n - 1 a : Optional[int] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) a : List[str] = 0 while count < prec: a : List[Any] = random.randint(2 , n - 1 ) a : Dict = bin_exp_mod(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if b != 1: a : Tuple = True for _ in range(SCREAMING_SNAKE_CASE_ ): if b == n - 1: a : str = False break a : Optional[Any] = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": a : Any = abs(int(input('''Enter bound : ''').strip())) print('''Here\'s the list of primes:''') print(''', '''.join(str(i) for i in range(n + 1) if is_prime_big(i)))
700
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
0
"""simple docstring""" import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 a : Any = get_tests_dir('''fixtures/dummy-config.json''') class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> List[Any]: a : Optional[Any] = 0 def __a ( self ) -> Dict: self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("transformers.models.auto" ) ) def __a ( self ) -> Optional[Any]: a : Dict = AutoConfig.from_pretrained("bert-base-uncased" ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> int: a : Tuple = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Any: a : List[str] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Dict: a : Any = AutoConfig.for_model("roberta" ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[Any]: with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. a : List[Any] = os.path.join(lowerCAmelCase__ , "fake-roberta" ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , "config.json" ) , "w" ) as f: f.write(json.dumps({} ) ) a : str = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertEqual(type(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __a ( self ) -> List[str]: try: AutoConfig.register("custom" , lowerCAmelCase__ ) # Wrong model type will raise an error with self.assertRaises(lowerCAmelCase__ ): AutoConfig.register("model" , lowerCAmelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCAmelCase__ ): AutoConfig.register("bert" , lowerCAmelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API a : List[str] = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__ ) a : Tuple = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def __a ( self ) -> Optional[int]: with self.assertRaisesRegex( lowerCAmelCase__ , "bert-base is not a local folder and is not a valid model identifier" ): a : List[str] = AutoConfig.from_pretrained("bert-base" ) def __a ( self ) -> int: with self.assertRaisesRegex( lowerCAmelCase__ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): a : Any = AutoConfig.from_pretrained(lowerCAmelCase__ , revision="aaaaaa" ) def __a ( self ) -> Optional[int]: with self.assertRaisesRegex( lowerCAmelCase__ , "hf-internal-testing/no-config-test-repo does not appear to have a file named config.json." , ): a : Any = AutoConfig.from_pretrained("hf-internal-testing/no-config-test-repo" ) def __a ( self ) -> Any: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(lowerCAmelCase__ ): a : List[str] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCAmelCase__ ): a : str = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__ ) a : int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__ ) a : Dict = AutoConfig.from_pretrained(lowerCAmelCase__ , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(reloaded_config.__class__.__name__ , "NewModelConfig" ) def __a ( self ) -> Tuple: class __UpperCamelCase ( _snake_case ): lowerCamelCase : str ="""new-model""" try: AutoConfig.register("new-model" , lowerCAmelCase__ ) # If remote code is not set, the default is to use local a : Dict = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote code is disabled, we load the local one. a : List[str] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote is enabled, we load from the Hub a : int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
701
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
0
"""simple docstring""" import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration a : Any = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('beta', 'bias'), ('gamma', 'weight'), ('pegasus', 'model'), ] a : Union[str, Any] = [ ('.output.dense', '.fc2'), ('intermediate.LayerNorm', 'final_layer_norm'), ('intermediate.dense', 'fc1'), ] a : Tuple = ( INIT_COMMON + [ ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.out_proj'), ('attention.self', 'self_attn'), ('attention.encdec.LayerNorm', 'encoder_attn_layer_norm'), ('attention.encdec_output.dense', 'encoder_attn.out_proj'), ('attention.encdec', 'encoder_attn'), ('key', 'k_proj'), ('value', 'v_proj'), ('query', 'q_proj'), ('decoder.LayerNorm', 'decoder.layernorm_embedding'), ] + END_COMMON ) a : List[str] = ( INIT_COMMON + [ ('embeddings.word_embeddings', 'shared.weight'), ('embeddings.position_embeddings', 'embed_positions.weight'), ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.output'), ('attention.self', 'self_attn.self'), ('encoder.LayerNorm', 'encoder.layernorm_embedding'), ] + END_COMMON ) a : Optional[int] = [ 'encdec/key/bias', 'encdec/query/bias', 'encdec/value/bias', 'self/key/bias', 'self/query/bias', 'self/value/bias', 'encdec_output/dense/bias', 'attention/output/dense/bias', ] def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Tuple ) ->Any: '''simple docstring''' for tf_name, hf_name in patterns: a : Tuple = k.replace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return k def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Dict ) ->BigBirdPegasusForConditionalGeneration: '''simple docstring''' a : Optional[int] = BigBirdPegasusConfig(**_SCREAMING_SNAKE_CASE ) a : Union[str, Any] = BigBirdPegasusForConditionalGeneration(_SCREAMING_SNAKE_CASE ) a : int = torch_model.state_dict() a : str = {} # separating decoder weights a : Optional[int] = {k: tf_weights[k] for k in tf_weights if k.startswith("pegasus/decoder" )} a : Tuple = {k: tf_weights[k] for k in tf_weights if not k.startswith("pegasus/decoder" )} for k, v in tqdm(decoder_weights.items() , "tf -> hf conversion" ): a : Optional[Any] = [k.endswith(_SCREAMING_SNAKE_CASE ) for ending in KEYS_TO_IGNORE] if any(_SCREAMING_SNAKE_CASE ): continue a : Tuple = DECODER_PATTERNS a : Optional[int] = rename_state_dict_key(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if new_k not in state_dict: raise ValueError(F"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ["dense", "query", "key", "value"] ): a : Optional[int] = v.T a : List[str] = torch.from_numpy(_SCREAMING_SNAKE_CASE ) assert v.shape == state_dict[new_k].shape, F"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" for k, v in tqdm(remaining_weights.items() , "tf -> hf conversion" ): a : Union[str, Any] = [k.endswith(_SCREAMING_SNAKE_CASE ) for ending in KEYS_TO_IGNORE] if any(_SCREAMING_SNAKE_CASE ): continue a : int = REMAINING_PATTERNS a : Tuple = rename_state_dict_key(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(F"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ["dense", "query", "key", "value"] ): a : Tuple = v.T a : List[Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, F"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" a : Any = mapping["model.embed_positions.weight"] a : Union[str, Any] = mapping.pop("model.embed_positions.weight" ) a, a : Optional[int] = torch_model.load_state_dict(_SCREAMING_SNAKE_CASE , strict=_SCREAMING_SNAKE_CASE ) a : List[str] = [ k for k in missing if k not in [ "final_logits_bias", "model.encoder.embed_tokens.weight", "model.decoder.embed_tokens.weight", "lm_head.weight", ] ] assert unexpected_missing == [], F"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], F"""no matches found for the following tf keys {extra}""" return torch_model def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->Dict: '''simple docstring''' a : Any = tf.train.list_variables(_SCREAMING_SNAKE_CASE ) a : List[Any] = {} a : Optional[Any] = ["global_step"] for name, shape in tqdm(_SCREAMING_SNAKE_CASE , desc="converting tf checkpoint to dict" ): a : Any = any(pat in name for pat in ignore_name ) if skip_key: continue a : str = tf.train.load_variable(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a : Dict = array return tf_weights def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Union[str, Any] , _lowercase : Tuple ) ->List[str]: '''simple docstring''' a : List[str] = get_tf_weights_as_numpy(_SCREAMING_SNAKE_CASE ) a : Optional[Any] = convert_bigbird_pegasus(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) torch_model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": a : List[str] = argparse.ArgumentParser() parser.add_argument('--tf_ckpt_path', type=str, help='passed to tf.train.list_variables') parser.add_argument('--save_dir', default=None, type=str, help='Path to the output PyTorch model.') a : Union[str, Any] = parser.parse_args() a : Dict = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
702
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
0
"""simple docstring""" from __future__ import annotations from typing import Generic, TypeVar a : Union[str, Any] = TypeVar('''T''') class __UpperCamelCase ( Generic[T] ): def __init__( self , lowerCAmelCase__ ) -> None: a : List[Any] = data a : Dict = self a : Union[str, Any] = 0 class __UpperCamelCase ( Generic[T] ): def __init__( self ) -> None: a : List[str] = {} def __a ( self , lowerCAmelCase__ ) -> None: a : Union[str, Any] = DisjointSetTreeNode(A_ ) def __a ( self , lowerCAmelCase__ ) -> DisjointSetTreeNode[T]: a : Optional[Any] = self.map[data] if elem_ref != elem_ref.parent: a : Dict = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: # helper function for union operation if nodea.rank > nodea.rank: a : Optional[Any] = nodea else: a : str = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: self.link(self.find_set(A_ ) , self.find_set(A_ ) ) class __UpperCamelCase ( Generic[T] ): def __init__( self ) -> None: a : List[Any] = {} def __a ( self , lowerCAmelCase__ ) -> None: # add a node ONLY if its not present in the graph if node not in self.connections: a : Any = {} def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: self.add_node(A_ ) self.add_node(A_ ) a : List[str] = weight a : str = weight def __a ( self ) -> GraphUndirectedWeighted[T]: a : List[Any] = [] a : Optional[Any] = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda lowerCAmelCase__ : x[2] ) # creating the disjoint set a : str = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(A_ ) # MST generation a : Optional[Any] = 0 a : Dict = 0 a : Union[str, Any] = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: a, a, a : Dict = edges[index] index += 1 a : str = disjoint_set.find_set(A_ ) a : Dict = disjoint_set.find_set(A_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(A_ , A_ , A_ ) disjoint_set.union(A_ , A_ ) return graph
703
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
0
"""simple docstring""" import math from collections.abc import Iterator from itertools import takewhile def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_lowercase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _SCREAMING_SNAKE_CASE ( ) ->Iterator[int]: '''simple docstring''' a : List[Any] = 2 while True: if is_prime(_lowercase ): yield num num += 1 def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] = 200_0000 ) ->int: '''simple docstring''' return sum(takewhile(lambda _lowercase : x < n , prime_generator() ) ) if __name__ == "__main__": print(F'''{solution() = }''')
704
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
0
"""simple docstring""" import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser a : Union[str, Any] = re.compile(R'''\s+''') def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->List[str]: '''simple docstring''' return {"hash": hashlib.mda(re.sub(_A , "" , example["content"] ).encode("utf-8" ) ).hexdigest()} def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->List[Any]: '''simple docstring''' a : str = [len(_A ) for line in example["content"].splitlines()] return {"line_mean": np.mean(_A ), "line_max": max(_A )} def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->List[Any]: '''simple docstring''' a : int = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] ) ->Tuple: '''simple docstring''' if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Dict=5 ) ->Union[str, Any]: '''simple docstring''' a : List[str] = ["auto-generated", "autogenerated", "automatically generated"] a : Tuple = example["content"].splitlines() for _, line in zip(range(_A ) , _A ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[int]=5 , _lowercase : Optional[int]=0.05 ) ->List[str]: '''simple docstring''' a : Optional[int] = ["unit tests", "test file", "configuration file"] a : int = example["content"].splitlines() a : List[Any] = 0 a : Optional[int] = 0 # first test for _, line in zip(range(_A ) , _A ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test a : Any = example["content"].count("\n" ) a : List[Any] = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->Dict: '''simple docstring''' a : List[Any] = ["def ", "class ", "for ", "while "] a : str = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : List[str]=4 ) ->int: '''simple docstring''' a : int = example["content"].splitlines() a : List[str] = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' a : Tuple = tokenizer(example["content"] , truncation=_A )["input_ids"] a : List[str] = len(example["content"] ) / len(_A ) return {"ratio": ratio} def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->List[Any]: '''simple docstring''' a : List[Any] = {} results.update(get_hash(_A ) ) results.update(line_stats(_A ) ) results.update(alpha_stats(_A ) ) results.update(char_token_ratio(_A ) ) results.update(is_autogenerated(_A ) ) results.update(is_config_or_test(_A ) ) results.update(has_no_keywords(_A ) ) results.update(has_few_assignments(_A ) ) return results def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Dict , _lowercase : int ) ->List[Any]: '''simple docstring''' if not check_uniques(_A , _A ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->int: '''simple docstring''' with open(_A , "rb" ) as f_in: with gzip.open(str(_A ) + ".gz" , "wb" , compresslevel=6 ) as f_out: shutil.copyfileobj(_A , _A ) os.unlink(_A ) # Settings a : str = HfArgumentParser(PreprocessingArguments) a : List[str] = parser.parse_args() if args.num_workers is None: a : Union[str, Any] = multiprocessing.cpu_count() a : int = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset a : str = time.time() a : str = load_dataset(args.dataset_name, split='''train''') print(F'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing a : List[str] = time.time() a : Dict = ds.map(preprocess, num_proc=args.num_workers) print(F'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes a : List[str] = set(ds.unique('''hash''')) a : int = len(uniques) / len(ds) print(F'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics a : List[str] = time.time() a : List[Any] = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'''Time to filter dataset: {time.time()-t_start:.2f}''') print(F'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: a : str = time.time() a : List[str] = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(F'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file a : Optional[int] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) a : str = output_dir / "data" data_dir.mkdir(exist_ok=True) a : Optional[Any] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): a : Any = str(data_dir / F'''file-{file_number+1:012}.json''') a : List[str] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'''Time to save dataset: {time.time()-t_start:.2f}''')
705
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
0
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> int: a : str = '| <pad> <unk> <s> </s> a b c d e f g h i j k'.split() a : Dict = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) a : str = { 'unk_token': '<unk>', 'bos_token': '<s>', 'eos_token': '</s>', } a : str = { 'feature_size': 1, 'padding_value': 0.0, 'sampling_rate': 1_6000, 'return_attention_mask': False, 'do_normalize': True, } a : List[str] = tempfile.mkdtemp() a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) a : Dict = os.path.join(self.tmpdirname , UpperCAmelCase_ ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + "\n" ) with open(self.feature_extraction_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + "\n" ) # load decoder from hub a : List[Any] = 'hf-internal-testing/ngram-beam-search-decoder' def __a ( self , **lowerCAmelCase__ ) -> List[Any]: a : Dict = self.add_kwargs_tokens_map.copy() kwargs.update(UpperCAmelCase_ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def __a ( self , **lowerCAmelCase__ ) -> Dict: return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def __a ( self , **lowerCAmelCase__ ) -> List[str]: return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **UpperCAmelCase_ ) def __a ( self ) -> int: shutil.rmtree(self.tmpdirname ) def __a ( self ) -> Union[str, Any]: a : List[Any] = self.get_tokenizer() a : str = self.get_feature_extractor() a : Union[str, Any] = self.get_decoder() a : int = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) processor.save_pretrained(self.tmpdirname ) a : int = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase_ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase_ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , UpperCAmelCase_ ) def __a ( self ) -> Union[str, Any]: a : Any = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match a : Tuple = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def __a ( self ) -> str: a : Union[str, Any] = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["xx"] ) with self.assertRaisesRegex(UpperCAmelCase_ , "include" ): WavaVecaProcessorWithLM( tokenizer=UpperCAmelCase_ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def __a ( self ) -> List[str]: a : Tuple = self.get_feature_extractor() a : Union[str, Any] = self.get_tokenizer() a : Tuple = self.get_decoder() a : Dict = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : Optional[int] = floats_list((3, 1000) ) a : List[Any] = feature_extractor(UpperCAmelCase_ , return_tensors="np" ) a : str = processor(UpperCAmelCase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __a ( self ) -> Optional[Any]: a : Optional[int] = self.get_feature_extractor() a : Union[str, Any] = self.get_tokenizer() a : str = self.get_decoder() a : Any = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : Any = 'This is a test string' a : List[Any] = processor(text=UpperCAmelCase_ ) a : str = tokenizer(UpperCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __a ( self , lowerCAmelCase__=(2, 10, 16) , lowerCAmelCase__=77 ) -> str: np.random.seed(UpperCAmelCase_ ) return np.random.rand(*UpperCAmelCase_ ) def __a ( self ) -> int: a : List[str] = self.get_feature_extractor() a : str = self.get_tokenizer() a : str = self.get_decoder() a : Any = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : Any = self._get_dummy_logits(shape=(10, 16) , seed=13 ) a : List[str] = processor.decode(UpperCAmelCase_ ) a : str = decoder.decode_beams(UpperCAmelCase_ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("</s> <s> </s>" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["fork"], ["spawn"]] ) def __a ( self , lowerCAmelCase__ ) -> Tuple: a : Union[str, Any] = self.get_feature_extractor() a : List[Any] = self.get_tokenizer() a : str = self.get_decoder() a : List[str] = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : Tuple = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: a : Optional[Any] = processor.batch_decode(UpperCAmelCase_ ) else: with get_context(UpperCAmelCase_ ).Pool() as pool: a : Optional[int] = processor.batch_decode(UpperCAmelCase_ , UpperCAmelCase_ ) a : Optional[Any] = list(UpperCAmelCase_ ) with get_context("fork" ).Pool() as p: a : Optional[Any] = decoder.decode_beams_batch(UpperCAmelCase_ , UpperCAmelCase_ ) a : Optional[Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(UpperCAmelCase_ , decoded_processor.text ) self.assertListEqual(["<s> <s> </s>", "<s> <s> <s>"] , decoded_processor.text ) self.assertListEqual(UpperCAmelCase_ , decoded_processor.logit_score ) self.assertListEqual(UpperCAmelCase_ , decoded_processor.lm_score ) def __a ( self ) -> Any: a : Dict = self.get_feature_extractor() a : List[Any] = self.get_tokenizer() a : str = self.get_decoder() a : Dict = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : Any = self._get_dummy_logits() a : Optional[Any] = 15 a : Optional[int] = -20.0 a : int = -4.0 a : List[str] = processor.batch_decode( UpperCAmelCase_ , beam_width=UpperCAmelCase_ , beam_prune_logp=UpperCAmelCase_ , token_min_logp=UpperCAmelCase_ , ) a : Any = decoded_processor_out.text a : List[Any] = list(UpperCAmelCase_ ) with get_context("fork" ).Pool() as pool: a : Any = decoder.decode_beams_batch( UpperCAmelCase_ , UpperCAmelCase_ , beam_width=UpperCAmelCase_ , beam_prune_logp=UpperCAmelCase_ , token_min_logp=UpperCAmelCase_ , ) a : List[str] = [d[0][0] for d in decoded_decoder_out] a : Union[str, Any] = [d[0][2] for d in decoded_decoder_out] a : Union[str, Any] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertListEqual(["</s> <s> <s>", "<s> <s> <s>"] , UpperCAmelCase_ ) self.assertTrue(np.array_equal(UpperCAmelCase_ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , UpperCAmelCase_ , atol=1E-3 ) ) self.assertTrue(np.array_equal(UpperCAmelCase_ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , UpperCAmelCase_ , atol=1E-3 ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = self.get_feature_extractor() a : Optional[int] = self.get_tokenizer() a : List[Any] = self.get_decoder() a : List[str] = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) a : int = self._get_dummy_logits() a : int = 2.0 a : List[Any] = 5.0 a : Union[str, Any] = -20.0 a : Any = True a : Tuple = processor.batch_decode( UpperCAmelCase_ , alpha=UpperCAmelCase_ , beta=UpperCAmelCase_ , unk_score_offset=UpperCAmelCase_ , lm_score_boundary=UpperCAmelCase_ , ) a : List[str] = decoded_processor_out.text a : Optional[int] = list(UpperCAmelCase_ ) decoder.reset_params( alpha=UpperCAmelCase_ , beta=UpperCAmelCase_ , unk_score_offset=UpperCAmelCase_ , lm_score_boundary=UpperCAmelCase_ , ) with get_context("fork" ).Pool() as pool: a : List[str] = decoder.decode_beams_batch( UpperCAmelCase_ , UpperCAmelCase_ , ) a : str = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertListEqual(["<s> </s> <s> </s> </s>", "</s> </s> <s> </s> </s>"] , UpperCAmelCase_ ) a : Any = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , UpperCAmelCase_ ) def __a ( self ) -> Union[str, Any]: a : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) a : List[Any] = processor.decoder.model_container[processor.decoder._model_key] a : Any = Path(language_model._kenlm_model.path.decode("utf-8" ) ).parent.parent.absolute() a : List[str] = os.listdir(UpperCAmelCase_ ) a : Any = ['alphabet.json', 'language_model'] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __a ( self ) -> List[Any]: a : List[str] = snapshot_download("hf-internal-testing/processor_with_lm" ) a : str = WavaVecaProcessorWithLM.from_pretrained(UpperCAmelCase_ ) a : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] a : Optional[int] = Path(language_model._kenlm_model.path.decode("utf-8" ) ).parent.parent.absolute() a : str = os.listdir(UpperCAmelCase_ ) a : str = os.listdir(UpperCAmelCase_ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __a ( self ) -> List[Any]: a : int = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) a : Optional[int] = AutoProcessor.from_pretrained("hf-internal-testing/processor_with_lm" ) a : List[str] = floats_list((3, 1000) ) a : Union[str, Any] = processor_wavaveca(UpperCAmelCase_ , return_tensors="np" ) a : Dict = processor_auto(UpperCAmelCase_ , return_tensors="np" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) a : Any = self._get_dummy_logits() a : int = processor_wavaveca.batch_decode(UpperCAmelCase_ ) a : Union[str, Any] = processor_auto.batch_decode(UpperCAmelCase_ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def __a ( self ) -> List[Any]: a : Dict = self.get_feature_extractor() a : Union[str, Any] = self.get_tokenizer() a : Optional[Any] = self.get_decoder() a : List[str] = WavaVecaProcessorWithLM(tokenizer=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: a : Tuple = [d[key] for d in offsets] return retrieved_list def __a ( self ) -> Tuple: a : List[Any] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) a : Dict = self._get_dummy_logits()[0] a : Any = processor.decode(UpperCAmelCase_ , output_word_offsets=UpperCAmelCase_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("text" in outputs ) self.assertTrue("word_offsets" in outputs ) self.assertTrue(isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertEqual(" ".join(self.get_from_offsets(outputs["word_offsets"] , "word" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "word" ) , ["<s>", "<s>", "</s>"] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "start_offset" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "end_offset" ) , [1, 3, 5] ) def __a ( self ) -> Optional[Any]: a : List[str] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) a : Dict = self._get_dummy_logits() a : Optional[int] = processor.batch_decode(UpperCAmelCase_ , output_word_offsets=UpperCAmelCase_ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("text" in outputs ) self.assertTrue("word_offsets" in outputs ) self.assertTrue(isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertListEqual( [" ".join(self.get_from_offsets(UpperCAmelCase_ , "word" ) ) for o in outputs["word_offsets"]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "word" ) , ["<s>", "<s>", "</s>"] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "start_offset" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "end_offset" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def __a ( self ) -> Optional[int]: import torch a : List[str] = load_dataset("common_voice" , "en" , split="train" , streaming=UpperCAmelCase_ ) a : int = ds.cast_column("audio" , datasets.Audio(sampling_rate=1_6000 ) ) a : Any = iter(UpperCAmelCase_ ) a : int = next(UpperCAmelCase_ ) a : List[Any] = AutoProcessor.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm" ) a : List[str] = WavaVecaForCTC.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train a : int = processor(sample["audio"]["array"] , return_tensors="pt" ).input_values with torch.no_grad(): a : Union[str, Any] = model(UpperCAmelCase_ ).logits.cpu().numpy() a : Optional[Any] = processor.decode(logits[0] , output_word_offsets=UpperCAmelCase_ ) a : Tuple = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate a : Union[str, Any] = [ { 'start_time': d['start_offset'] * time_offset, 'end_time': d['end_offset'] * time_offset, 'word': d['word'], } for d in output['word_offsets'] ] a : Union[str, Any] = 'WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL' # output words self.assertEqual(" ".join(self.get_from_offsets(UpperCAmelCase_ , "word" ) ) , UpperCAmelCase_ ) self.assertEqual(" ".join(self.get_from_offsets(UpperCAmelCase_ , "word" ) ) , output.text ) # output times a : Tuple = torch.tensor(self.get_from_offsets(UpperCAmelCase_ , "start_time" ) ) a : Tuple = torch.tensor(self.get_from_offsets(UpperCAmelCase_ , "end_time" ) ) # fmt: off a : Any = torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) a : Union[str, Any] = torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(UpperCAmelCase_ , UpperCAmelCase_ , atol=0.01 ) ) self.assertTrue(torch.allclose(UpperCAmelCase_ , UpperCAmelCase_ , atol=0.01 ) )
706
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
0
a : Union[str, Any] = { 0: "0", 1: "1", 2: "2", 3: "3", 4: "4", 5: "5", 6: "6", 7: "7", 8: "8", 9: "9", 10: "a", 11: "b", 12: "c", 13: "d", 14: "e", 15: "f", } def _SCREAMING_SNAKE_CASE ( _lowercase : float ) ->str: '''simple docstring''' assert type(_lowercase ) in (int, float) and decimal == int(_lowercase ) a : Tuple = int(_lowercase ) a : int = '' a : Tuple = False if decimal < 0: a : Optional[Any] = True decimal *= -1 while decimal > 0: a : Dict = divmod(_lowercase , 16 ) a : Tuple = values[remainder] + hexadecimal a : Optional[Any] = '0x' + hexadecimal if negative: a : str = '-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
707
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
0
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 ) ->str: '''simple docstring''' if not isinstance(__A , __A ) or n < 0: raise ValueError("Invalid input" ) a : str = 10**n a : Optional[int] = 2_8433 * (pow(2 , 783_0457 , __A )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F'''{solution(10) = }''')
708
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
0
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def _SCREAMING_SNAKE_CASE ( ) ->Union[str, Any]: '''simple docstring''' a : Tuple = HfArgumentParser(__lowerCAmelCase ) a : List[str] = parser.parse_args_into_dataclasses()[0] a : Optional[int] = TensorFlowBenchmark(args=__lowerCAmelCase ) try: a : Optional[int] = parser.parse_args_into_dataclasses()[0] except ValueError as e: a : str = "Arg --no_{0} is no longer used, please use --no-{0} instead." a : Optional[Any] = " ".join(str(__lowerCAmelCase ).split(" " )[:-1] ) a : List[Any] = "" a : int = eval(str(__lowerCAmelCase ).split(" " )[-1] ) a : Tuple = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: a : str = full_error_msg + begin_error_msg + str(__lowerCAmelCase ) raise ValueError(__lowerCAmelCase ) benchmark.run() if __name__ == "__main__": main()
709
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
0
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer a : Optional[int] = logging.get_logger(__name__) a : int = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} a : Optional[int] = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } a : List[Any] = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } a : List[str] = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } a : Optional[Any] = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } a : Dict = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } a : Optional[int] = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } a : List[Any] = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } a : Any = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } a : str = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : List[Any] =CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Tuple =CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] =CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION lowerCamelCase : List[str] =DPRContextEncoderTokenizer class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[int] =VOCAB_FILES_NAMES lowerCamelCase : Union[str, Any] =QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Dict =QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Union[str, Any] =QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION lowerCamelCase : List[Any] =DPRQuestionEncoderTokenizer a : List[Any] = collections.namedtuple( '''DPRSpanPrediction''', ['''span_score''', '''relevance_score''', '''doc_id''', '''start_index''', '''end_index''', '''text'''] ) a : List[str] = collections.namedtuple('''DPRReaderOutput''', ['''start_logits''', '''end_logits''', '''relevance_logits''']) a : List[str] = R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(a__ ) class __UpperCamelCase : def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Dict: if titles is None and texts is None: return super().__call__( UpperCamelCase_ , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=UpperCamelCase_ , return_tensors=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) elif titles is None or texts is None: a : Union[str, Any] = titles if texts is None else texts return super().__call__( UpperCamelCase_ , UpperCamelCase_ , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=UpperCamelCase_ , return_tensors=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) a : Union[str, Any] = titles if not isinstance(UpperCamelCase_ , UpperCamelCase_ ) else [titles] a : List[Any] = texts if not isinstance(UpperCamelCase_ , UpperCamelCase_ ) else [texts] a : Union[str, Any] = len(UpperCamelCase_ ) a : Dict = questions if not isinstance(UpperCamelCase_ , UpperCamelCase_ ) else [questions] * n_passages assert len(UpperCamelCase_ ) == len( UpperCamelCase_ ), f"""There should be as many titles than texts but got {len(UpperCamelCase_ )} titles and {len(UpperCamelCase_ )} texts.""" a : Dict = super().__call__(UpperCamelCase_ , UpperCamelCase_ , padding=UpperCamelCase_ , truncation=UpperCamelCase_ )["input_ids"] a : str = super().__call__(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ , padding=UpperCamelCase_ , truncation=UpperCamelCase_ )["input_ids"] a : List[str] = { "input_ids": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(UpperCamelCase_ , UpperCamelCase_ ) ] } if return_attention_mask is not False: a : Optional[Any] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) a : str = attention_mask return self.pad(UpperCamelCase_ , padding=UpperCamelCase_ , max_length=UpperCamelCase_ , return_tensors=UpperCamelCase_ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 16 , lowerCAmelCase__ = 64 , lowerCAmelCase__ = 4 , ) -> int: a : Optional[Any] = reader_input["input_ids"] a : int = reader_output[:3] a : int = len(UpperCamelCase_ ) a : List[str] = sorted(range(UpperCamelCase_ ) , reverse=UpperCamelCase_ , key=relevance_logits.__getitem__ ) a : List[DPRReaderOutput] = [] for doc_id in sorted_docs: a : Optional[Any] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence a : Dict = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: a : List[Any] = sequence_ids.index(self.pad_token_id ) else: a : str = len(UpperCamelCase_ ) a : Optional[Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=UpperCamelCase_ , top_spans=UpperCamelCase_ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=UpperCamelCase_ , start_index=UpperCamelCase_ , end_index=UpperCamelCase_ , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(UpperCamelCase_ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> Optional[Any]: a : List[Any] = [] for start_index, start_score in enumerate(UpperCamelCase_ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) a : List[Any] = sorted(UpperCamelCase_ , key=lambda lowerCAmelCase__ : x[1] , reverse=UpperCamelCase_ ) a : Optional[int] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f"""Wrong span indices: [{start_index}:{end_index}]""" a : Tuple = end_index - start_index + 1 assert length <= max_answer_length, f"""Span is too long: {length} > {max_answer_length}""" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(UpperCamelCase_ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(a__ ) class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : int =VOCAB_FILES_NAMES lowerCamelCase : int =READER_PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Dict =READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =READER_PRETRAINED_INIT_CONFIGURATION lowerCamelCase : int =["input_ids", "attention_mask"] lowerCamelCase : str =DPRReaderTokenizer
710
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
0
"""simple docstring""" import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node a : Dict = 4 a : List[str] = 3 class __UpperCamelCase ( a__ ): pass def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->Optional[int]: '''simple docstring''' for shard in shards: for i in range(_lowerCAmelCase ): yield {"i": i, "shard": shard} def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : str = int(os.environ["RANK"] ) a : List[str] = int(os.environ["WORLD_SIZE"] ) a : Optional[Any] = ArgumentParser() parser.add_argument("--streaming" , type=_lowerCAmelCase ) parser.add_argument("--local_rank" , type=_lowerCAmelCase ) parser.add_argument("--num_workers" , type=_lowerCAmelCase , default=0 ) a : Optional[int] = parser.parse_args() a : Tuple = args.streaming a : Any = args.num_workers a : Tuple = {"shards": [F"""shard_{shard_idx}""" for shard_idx in range(_lowerCAmelCase )]} a : int = IterableDataset.from_generator(_lowerCAmelCase , gen_kwargs=_lowerCAmelCase ) if not streaming: a : List[str] = Dataset.from_list(list(_lowerCAmelCase ) ) a : Optional[Any] = split_dataset_by_node(_lowerCAmelCase , rank=_lowerCAmelCase , world_size=_lowerCAmelCase ) a : List[str] = torch.utils.data.DataLoader(_lowerCAmelCase , num_workers=_lowerCAmelCase ) a : Union[str, Any] = NUM_SHARDS * NUM_ITEMS_PER_SHARD a : Optional[int] = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) a : Union[str, Any] = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(F"""local_size {local_size} != expected_local_size {expected_local_size}""" ) if __name__ == "__main__": main()
711
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
0
"""simple docstring""" from __future__ import annotations import time a : Optional[int] = list[tuple[int, int]] a : Any = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] a : Optional[int] = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: a : Dict = pos_x a : Tuple = pos_y a : Optional[int] = (pos_y, pos_x) a : Optional[int] = goal_x a : List[Any] = goal_y a : Dict = parent class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: a : Any = Node(start[1] , start[0] , goal[1] , goal[0] , _a ) a : Tuple = Node(goal[1] , goal[0] , goal[1] , goal[0] , _a ) a : Union[str, Any] = [self.start] a : Union[str, Any] = False def __a ( self ) -> Path | None: while self.node_queue: a : str = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: a : Tuple = True return self.retrace_path(_a ) a : Dict = self.get_successors(_a ) for node in successors: self.node_queue.append(_a ) if not self.reached: return [self.start.pos] return None def __a ( self , lowerCAmelCase__ ) -> list[Node]: a : Union[str, Any] = [] for action in delta: a : Union[str, Any] = parent.pos_x + action[1] a : Optional[Any] = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(_a ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(_a , _a , self.target.pos_y , self.target.pos_x , _a ) ) return successors def __a ( self , lowerCAmelCase__ ) -> Path: a : Union[str, Any] = node a : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) a : Optional[int] = current_node.parent path.reverse() return path class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: a : List[Any] = BreadthFirstSearch(_a , _a ) a : Union[str, Any] = BreadthFirstSearch(_a , _a ) a : List[str] = False def __a ( self ) -> Path | None: while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: a : str = self.fwd_bfs.node_queue.pop(0 ) a : str = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: a : List[str] = True return self.retrace_bidirectional_path( _a , _a ) a : Union[str, Any] = current_bwd_node a : Tuple = current_fwd_node a : List[Any] = { self.fwd_bfs: self.fwd_bfs.get_successors(_a ), self.bwd_bfs: self.bwd_bfs.get_successors(_a ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(_a ) if not self.reached: return [self.fwd_bfs.start.pos] return None def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Path: a : int = self.fwd_bfs.retrace_path(_a ) a : List[Any] = self.bwd_bfs.retrace_path(_a ) bwd_path.pop() bwd_path.reverse() a : Optional[int] = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() a : Union[str, Any] = (0, 0) a : Optional[int] = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) a : Dict = time.time() a : Any = BreadthFirstSearch(init, goal) a : str = bfs.search() a : Union[str, Any] = time.time() - start_bfs_time print('''Unidirectional BFS computation time : ''', bfs_time) a : Union[str, Any] = time.time() a : int = BidirectionalBreadthFirstSearch(init, goal) a : int = bd_bfs.search() a : int = time.time() - start_bd_bfs_time print('''Bidirectional BFS computation time : ''', bd_bfs_time)
712
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
0
"""simple docstring""" import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : int , _lowercase : List[Any] , _lowercase : List[Any] , _lowercase : Union[str, Any] ): '''simple docstring''' a : List[str] = StableDiffusionPipeline.from_pretrained(__UpperCAmelCase , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors a : List[Any] = load_file(__UpperCAmelCase ) a : Any = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: a : str = key.split("." )[0].split(LORA_PREFIX_TEXT_ENCODER + "_" )[-1].split("_" ) a : List[str] = pipeline.text_encoder else: a : Optional[int] = key.split("." )[0].split(LORA_PREFIX_UNET + "_" )[-1].split("_" ) a : Any = pipeline.unet # find the target layer a : List[Any] = layer_infos.pop(0 ) while len(__UpperCAmelCase ) > -1: try: a : Optional[Any] = curr_layer.__getattr__(__UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: a : Any = layer_infos.pop(0 ) elif len(__UpperCAmelCase ) == 0: break except Exception: if len(__UpperCAmelCase ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: a : Optional[int] = layer_infos.pop(0 ) a : Union[str, Any] = [] if "lora_down" in key: pair_keys.append(key.replace("lora_down" , "lora_up" ) ) pair_keys.append(__UpperCAmelCase ) else: pair_keys.append(__UpperCAmelCase ) pair_keys.append(key.replace("lora_up" , "lora_down" ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: a : Optional[Any] = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) a : Tuple = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(__UpperCAmelCase , __UpperCAmelCase ).unsqueeze(2 ).unsqueeze(3 ) else: a : Optional[Any] = state_dict[pair_keys[0]].to(torch.floataa ) a : Optional[Any] = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(__UpperCAmelCase , __UpperCAmelCase ) # update visited list for item in pair_keys: visited.append(__UpperCAmelCase ) return pipeline if __name__ == "__main__": a : List[Any] = argparse.ArgumentParser() parser.add_argument( '''--base_model_path''', default=None, type=str, required=True, help='''Path to the base model in diffusers format.''' ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--lora_prefix_unet''', default='''lora_unet''', type=str, help='''The prefix of UNet weight in safetensors''' ) parser.add_argument( '''--lora_prefix_text_encoder''', default='''lora_te''', type=str, help='''The prefix of text encoder weight in safetensors''', ) parser.add_argument('''--alpha''', default=0.75, type=float, help='''The merging ratio in W = W0 + alpha * deltaW''') parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''' ) parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') a : Any = parser.parse_args() a : Any = args.base_model_path a : Dict = args.checkpoint_path a : Optional[int] = args.dump_path a : List[str] = args.lora_prefix_unet a : str = args.lora_prefix_text_encoder a : List[str] = args.alpha a : Dict = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) a : Dict = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
713
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
0
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() a : Union[str, Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : Union[str, Any]=False ) ->str: '''simple docstring''' a : Dict = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("head" ): a : Union[str, Any] = "segformer.encoder." + key if key.startswith("backbone" ): a : Union[str, Any] = key.replace("backbone" , "segformer.encoder" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 a : List[Any] = key[key.find("patch_embed" ) + len("patch_embed" )] a : str = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(lowerCAmelCase__ )-1}""" ) if "norm" in key: a : Tuple = key.replace("norm" , "layer_norm" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 a : List[Any] = key[key.find("segformer.encoder.layer_norm" ) + len("segformer.encoder.layer_norm" )] a : Optional[int] = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(lowerCAmelCase__ )-1}""" ) if "layer_norm1" in key: a : Dict = key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: a : List[str] = key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 a : Dict = key[key.find("block" ) + len("block" )] a : Any = key.replace(F"""block{idx}""" , F"""block.{int(lowerCAmelCase__ )-1}""" ) if "attn.q" in key: a : int = key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: a : Optional[int] = key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: a : Dict = key.replace("attn" , "attention.self" ) if "fc1" in key: a : str = key.replace("fc1" , "dense1" ) if "fc2" in key: a : Tuple = key.replace("fc2" , "dense2" ) if "linear_pred" in key: a : Union[str, Any] = key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: a : int = key.replace("linear_fuse.conv" , "linear_fuse" ) a : Any = key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 a : int = key[key.find("linear_c" ) + len("linear_c" )] a : Union[str, Any] = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(lowerCAmelCase__ )-1}""" ) if key.startswith("head" ): a : Optional[int] = key.replace("head" , "classifier" ) a : Any = value return new_state_dict def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Union[str, Any] ) ->Optional[int]: '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) a : List[str] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) a : Any = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict a : Dict = kv_weight[ : config.hidden_sizes[i], : ] a : List[Any] = kv_bias[: config.hidden_sizes[i]] a : List[str] = kv_weight[ config.hidden_sizes[i] :, : ] a : Dict = kv_bias[ config.hidden_sizes[i] : ] def _SCREAMING_SNAKE_CASE ( ) ->Union[str, Any]: '''simple docstring''' a : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" a : Any = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ) return image @torch.no_grad() def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : str , _lowercase : Dict ) ->Optional[Any]: '''simple docstring''' a : List[Any] = SegformerConfig() a : Dict = False # set attributes based on model_name a : Optional[Any] = "huggingface/label-files" if "segformer" in model_name: a : Optional[int] = model_name[len("segformer." ) : len("segformer." ) + 2] if "ade" in model_name: a : int = 150 a : List[Any] = "ade20k-id2label.json" a : Dict = (1, 150, 128, 128) elif "city" in model_name: a : List[Any] = 19 a : Optional[int] = "cityscapes-id2label.json" a : Union[str, Any] = (1, 19, 128, 128) else: raise ValueError(F"""Model {model_name} not supported""" ) elif "mit" in model_name: a : Optional[Any] = True a : Optional[int] = model_name[4:6] a : List[str] = 1000 a : Union[str, Any] = "imagenet-1k-id2label.json" a : Optional[Any] = (1, 1000) else: raise ValueError(F"""Model {model_name} not supported""" ) # set config attributes a : int = json.load(open(hf_hub_download(lowerCAmelCase__ , lowerCAmelCase__ , repo_type="dataset" ) , "r" ) ) a : int = {int(lowerCAmelCase__ ): v for k, v in idalabel.items()} a : str = idalabel a : List[str] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": a : int = [64, 128, 320, 512] a : int = 256 elif size == "b2": a : str = [64, 128, 320, 512] a : List[Any] = 768 a : int = [3, 4, 6, 3] elif size == "b3": a : List[Any] = [64, 128, 320, 512] a : Any = 768 a : Optional[Any] = [3, 4, 18, 3] elif size == "b4": a : Optional[Any] = [64, 128, 320, 512] a : Tuple = 768 a : List[Any] = [3, 8, 27, 3] elif size == "b5": a : Tuple = [64, 128, 320, 512] a : Dict = 768 a : Optional[Any] = [3, 6, 40, 3] else: raise ValueError(F"""Size {size} not supported""" ) # load image processor (only resize + normalize) a : List[Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=lowerCAmelCase__ , align=lowerCAmelCase__ , do_random_crop=lowerCAmelCase__ ) # prepare image a : str = prepare_img() a : List[str] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict if encoder_only: a : str = torch.load(lowerCAmelCase__ , map_location=torch.device("cpu" ) ) else: a : Tuple = torch.load(lowerCAmelCase__ , map_location=torch.device("cpu" ) )["state_dict"] # rename keys a : Dict = rename_keys(lowerCAmelCase__ , encoder_only=lowerCAmelCase__ ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(lowerCAmelCase__ , lowerCAmelCase__ ) # create HuggingFace model and load state dict if encoder_only: a : Tuple = False a : Optional[int] = SegformerForImageClassification(lowerCAmelCase__ ) else: a : List[Any] = SegformerForSemanticSegmentation(lowerCAmelCase__ ) model.load_state_dict(lowerCAmelCase__ ) model.eval() # forward pass a : Tuple = model(lowerCAmelCase__ ) a : Optional[int] = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": a : List[str] = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": a : List[Any] = torch.tensor( [ [[-7.5820, -8.7231, -8.3215], [-8.0600, -10.3529, -10.0304], [-7.5208, -9.4103, -9.6239]], [[-12.6918, -13.8994, -13.7137], [-13.3196, -15.7523, -15.4789], [-12.9343, -14.8757, -14.9689]], [[-11.1911, -11.9421, -11.3243], [-11.3342, -13.6839, -13.3581], [-10.3909, -12.1832, -12.4858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": a : Tuple = torch.tensor( [ [[-11.8173, -14.3850, -16.3128], [-14.5648, -16.5804, -18.6568], [-14.7223, -15.7387, -18.4218]], [[-15.7290, -17.9171, -19.4423], [-18.3105, -19.9448, -21.4661], [-17.9296, -18.6497, -20.7910]], [[-15.0783, -17.0336, -18.2789], [-16.8771, -18.6870, -20.1612], [-16.2454, -17.1426, -19.5055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": a : Any = torch.tensor( [ [[-9.0878, -10.2081, -10.1891], [-9.3144, -10.7941, -10.9843], [-9.2294, -10.3855, -10.5704]], [[-12.2316, -13.9068, -13.6102], [-12.9161, -14.3702, -14.3235], [-12.5233, -13.7174, -13.7932]], [[-14.6275, -15.2490, -14.9727], [-14.3400, -15.9687, -16.2827], [-14.1484, -15.4033, -15.8937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": a : Optional[Any] = torch.tensor( [ [[-12.3144, -13.2447, -14.0802], [-13.3614, -14.5816, -15.6117], [-13.3340, -14.4433, -16.2219]], [[-19.2781, -20.4128, -20.7506], [-20.6153, -21.6566, -22.0998], [-19.9800, -21.0430, -22.1494]], [[-18.8739, -19.7804, -21.1834], [-20.1233, -21.6765, -23.2944], [-20.0315, -21.2641, -23.6944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": a : Tuple = torch.tensor( [ [[-9.5524, -12.0835, -11.7348], [-10.5229, -13.6446, -14.5662], [-9.5842, -12.8851, -13.9414]], [[-15.3432, -17.5323, -17.0818], [-16.3330, -18.9255, -19.2101], [-15.1340, -17.7848, -18.3971]], [[-12.6072, -14.9486, -14.6631], [-13.7629, -17.0907, -17.7745], [-12.7899, -16.1695, -17.1671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": a : Union[str, Any] = torch.tensor( [ [[-11.9295, -13.4057, -14.8106], [-13.3431, -14.8179, -15.3781], [-14.2836, -15.5942, -16.1588]], [[-11.4906, -12.8067, -13.6564], [-13.1189, -14.0500, -14.1543], [-13.8748, -14.5136, -14.8789]], [[0.5374, 0.1067, -0.4742], [0.1141, -0.2255, -0.7099], [-0.3000, -0.5924, -1.3105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": a : List[Any] = torch.tensor( [ [[-7.8217, -9.8767, -10.1717], [-9.4438, -10.9058, -11.4047], [-9.7939, -12.3495, -12.1079]], [[-7.1514, -9.5336, -10.0860], [-9.7776, -11.6822, -11.8439], [-10.1411, -12.7655, -12.8972]], [[0.3021, 0.0805, -0.2310], [-0.0328, -0.1605, -0.2714], [-0.1408, -0.5477, -0.6976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": a : Dict = torch.tensor( [ [ [-1.1_3_7_2E0_1, -1.2_7_8_7E0_1, -1.3_4_7_7E0_1], [-1.2_5_3_6E0_1, -1.4_1_9_4E0_1, -1.4_4_0_9E0_1], [-1.3_2_1_7E0_1, -1.4_8_8_8E0_1, -1.5_3_2_7E0_1], ], [ [-1.4_7_9_1E0_1, -1.7_1_2_2E0_1, -1.8_2_7_7E0_1], [-1.7_1_6_3E0_1, -1.9_1_9_2E0_1, -1.9_5_3_3E0_1], [-1.7_8_9_7E0_1, -1.9_9_9_1E0_1, -2.0_3_1_5E0_1], ], [ [7.6_7_2_3E-0_1, 4.1_9_2_1E-0_1, -7.7_8_7_8E-0_2], [4.7_7_7_2E-0_1, 9.5_5_5_7E-0_3, -2.8_0_8_2E-0_1], [3.6_0_3_2E-0_1, -2.4_8_2_6E-0_1, -5.1_1_6_8E-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": a : Dict = torch.tensor( [ [[-9.4959, -11.3087, -11.7479], [-11.0025, -12.6540, -12.3319], [-11.4064, -13.0487, -12.9905]], [[-9.8905, -11.3084, -12.0854], [-11.1726, -12.7698, -12.9583], [-11.5985, -13.3278, -14.1774]], [[0.2213, 0.0192, -0.2466], [-0.1731, -0.4213, -0.4874], [-0.3126, -0.6541, -1.1389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": a : Union[str, Any] = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": a : Tuple = torch.tensor( [ [[-16.0976, -16.4856, -17.3962], [-16.6234, -19.0342, -19.7685], [-16.0900, -18.0661, -19.1180]], [[-18.4750, -18.8488, -19.5074], [-19.4030, -22.1570, -22.5977], [-19.1191, -20.8486, -22.3783]], [[-4.5178, -5.5037, -6.5109], [-5.0884, -7.2174, -8.0334], [-4.4156, -5.8117, -7.2970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": a : str = torch.tensor( [ [[-14.2081, -14.4732, -14.1977], [-14.5867, -16.4423, -16.6356], [-13.4441, -14.9685, -16.8696]], [[-14.4576, -14.7073, -15.0451], [-15.0816, -17.6237, -17.9873], [-14.4213, -16.0199, -18.5992]], [[-4.7349, -4.9588, -5.0966], [-4.3210, -6.9325, -7.2591], [-3.4312, -4.7484, -7.1917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": a : Tuple = torch.tensor( [ [[-11.7737, -11.9526, -11.3273], [-13.6692, -14.4574, -13.8878], [-13.8937, -14.6924, -15.9345]], [[-14.6706, -14.5330, -14.1306], [-16.1502, -16.8180, -16.4269], [-16.8338, -17.8939, -20.1746]], [[1.0491, 0.8289, 1.0310], [1.1044, 0.5219, 0.8055], [1.0899, 0.6926, 0.5590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": a : List[Any] = torch.tensor( [ [[-12.5641, -13.4777, -13.0684], [-13.9587, -15.8983, -16.6557], [-13.3109, -15.7350, -16.3141]], [[-14.7074, -15.4352, -14.5944], [-16.6353, -18.1663, -18.6120], [-15.1702, -18.0329, -18.1547]], [[-1.7990, -2.0951, -1.7784], [-2.6397, -3.8245, -3.9686], [-1.5264, -2.8126, -2.9316]], ] ) else: a : int = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) image_processor.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": a : Any = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) a : Any = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
714
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __UpperCamelCase ( __lowerCAmelCase , unittest.TestCase ): lowerCamelCase : Tuple =KandinskyVaaControlnetImgaImgPipeline lowerCamelCase : Tuple =["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] lowerCamelCase : int =["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] lowerCamelCase : Optional[int] =[ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] lowerCamelCase : str =False @property def __a ( self ) -> int: return 32 @property def __a ( self ) -> Optional[Any]: return 32 @property def __a ( self ) -> int: return self.time_input_dim @property def __a ( self ) -> Union[str, Any]: return self.time_input_dim * 4 @property def __a ( self ) -> str: return 100 @property def __a ( self ) -> List[Any]: torch.manual_seed(0 ) a : Optional[Any] = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } a : List[Any] = UNetaDConditionModel(**_UpperCamelCase ) return model @property def __a ( self ) -> List[Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __a ( self ) -> str: torch.manual_seed(0 ) a : List[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self ) -> Tuple: a : Union[str, Any] = self.dummy_unet a : Dict = self.dummy_movq a : List[str] = { """num_train_timesteps""": 1000, """beta_schedule""": """linear""", """beta_start""": 0.00_085, """beta_end""": 0.012, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } a : Dict = DDIMScheduler(**_UpperCamelCase ) a : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> str: a : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) a : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _UpperCamelCase ) # create init_image a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) a : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] a : int = Image.fromarray(np.uinta(_UpperCamelCase ) ).convert("RGB" ).resize((256, 256) ) # create hint a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) if str(_UpperCamelCase ).startswith("mps" ): a : Any = torch.manual_seed(_UpperCamelCase ) else: a : Any = torch.Generator(device=_UpperCamelCase ).manual_seed(_UpperCamelCase ) a : List[Any] = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def __a ( self ) -> Dict: a : str = """cpu""" a : Tuple = self.get_dummy_components() a : Union[str, Any] = self.pipeline_class(**_UpperCamelCase ) a : str = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) a : List[str] = pipe(**self.get_dummy_inputs(_UpperCamelCase ) ) a : Any = output.images a : Dict = pipe( **self.get_dummy_inputs(_UpperCamelCase ) , return_dict=_UpperCamelCase , )[0] a : Optional[int] = image[0, -3:, -3:, -1] a : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) a : str = np.array( [0.54_985_034, 0.55_509_365, 0.52_561_504, 0.5_570_494, 0.5_593_818, 0.5_263_979, 0.50_285_643, 0.5_069_846, 0.51_196_736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Dict: super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> int: a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy" ) a : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) a : Optional[int] = init_image.resize((512, 512) ) a : Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png" ) a : int = torch.from_numpy(np.array(_UpperCamelCase ) ).float() / 255.0 a : Dict = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) a : Tuple = """A robot, 4k photo""" a : Optional[int] = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(_UpperCamelCase ) a : str = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa ) a : List[str] = pipeline.to(_UpperCamelCase ) pipeline.set_progress_bar_config(disable=_UpperCamelCase ) a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Union[str, Any] = pipe_prior( _UpperCamelCase , image=_UpperCamelCase , strength=0.85 , generator=_UpperCamelCase , negative_prompt="" , ).to_tuple() a : Optional[Any] = pipeline( image=_UpperCamelCase , image_embeds=_UpperCamelCase , negative_image_embeds=_UpperCamelCase , hint=_UpperCamelCase , generator=_UpperCamelCase , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(_UpperCamelCase , _UpperCamelCase )
715
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
0
"""simple docstring""" from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. a : List[Any] = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. a : Any = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. a : str = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : str ) ->int: '''simple docstring''' a : Optional[Any] = len([g for position, g in enumerate(_lowerCamelCase ) if g == main_target[position]] ) return (item, float(_lowerCamelCase )) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : str ) ->Any: '''simple docstring''' a : Tuple = random.randint(0 , len(_lowerCamelCase ) - 1 ) a : List[str] = parent_a[:random_slice] + parent_a[random_slice:] a : Optional[Any] = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Any ) ->Optional[int]: '''simple docstring''' a : Union[str, Any] = list(_lowerCamelCase ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: a : List[Any] = random.choice(_lowerCamelCase ) return "".join(_lowerCamelCase ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : List[str] , _lowercase : int , ) ->Optional[Any]: '''simple docstring''' a : Optional[Any] = [] # Generate more children proportionally to the fitness score. a : List[str] = int(parent_a[1] * 100 ) + 1 a : List[Any] = 10 if child_n >= 10 else child_n for _ in range(_lowerCamelCase ): a : Union[str, Any] = population_score[random.randint(0 , _lowerCamelCase )][0] a : int = crossover(parent_a[0] , _lowerCamelCase ) # Append new string to the population list. pop.append(mutate(_lowerCamelCase , _lowerCamelCase ) ) pop.append(mutate(_lowerCamelCase , _lowerCamelCase ) ) return pop def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : List[Any] , _lowercase : Dict = True ) ->int: '''simple docstring''' if N_POPULATION < N_SELECTED: a : Tuple = F"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(_lowerCamelCase ) # Verify that the target contains no genes besides the ones inside genes variable. a : Optional[int] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: a : Any = F"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(_lowerCamelCase ) # Generate random starting population. a : List[Any] = [] for _ in range(_lowerCamelCase ): population.append("".join([random.choice(_lowerCamelCase ) for i in range(len(_lowerCamelCase ) )] ) ) # Just some logs to know what the algorithms is doing. a : List[str] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_lowerCamelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. a : Any = [evaluate(_lowerCamelCase , _lowerCamelCase ) for item in population] # Check if there is a matching evolution. a : Tuple = sorted(_lowerCamelCase , key=lambda _lowercase : x[1] , reverse=_lowerCamelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F"""\nGeneration: {generation}""" F"""\nTotal Population:{total_population}""" F"""\nBest score: {population_score[0][1]}""" F"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. a : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_lowerCamelCase ) # Normalize population score to be between 0 and 1. a : int = [ (item, score / len(_lowerCamelCase )) for item, score in population_score ] # This is selection for i in range(_lowerCamelCase ): population.extend(select(population_score[int(_lowerCamelCase )] , _lowerCamelCase , _lowerCamelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_lowerCamelCase ) > N_POPULATION: break if __name__ == "__main__": a : Any = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) a : int = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) a , a , a : Union[str, Any] = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
716
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
0
"""simple docstring""" from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class __UpperCamelCase : lowerCamelCase : Optional[int] =42 lowerCamelCase : Union[str, Any] =42 class __UpperCamelCase : def __init__( self , lowerCAmelCase__ ) -> Tuple: a : list[list[Edge]] = [[] for _ in range(__lowerCamelCase )] a : int = size def __getitem__( self , lowerCAmelCase__ ) -> Iterator[Edge]: return iter(self._graph[vertex] ) @property def __a ( self ) -> Tuple: return self._size def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: if weight not in (0, 1): raise ValueError("Edge weight must be either 0 or 1." ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("Vertex indexes must be in [0; size)." ) self._graph[from_vertex].append(Edge(__lowerCamelCase , __lowerCamelCase ) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int | None: a : List[Any] = deque([start_vertex] ) a : list[int | None] = [None] * self.size a : str = 0 while queue: a : Union[str, Any] = queue.popleft() a : int = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: a : Tuple = current_distance + edge.weight a : str = distances[edge.destination_vertex] if ( isinstance(__lowerCamelCase , __lowerCamelCase ) and new_distance >= dest_vertex_distance ): continue a : Tuple = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("No path from start_vertex to finish_vertex." ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
717
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[Any] = logging.get_logger(__name__) a : Optional[int] = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class __UpperCamelCase ( _UpperCAmelCase ): lowerCamelCase : str ="""swinv2""" lowerCamelCase : Optional[int] ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , lowerCAmelCase__=224 , lowerCAmelCase__=4 , lowerCAmelCase__=3 , lowerCAmelCase__=96 , lowerCAmelCase__=[2, 2, 6, 2] , lowerCAmelCase__=[3, 6, 12, 24] , lowerCAmelCase__=7 , lowerCAmelCase__=4.0 , lowerCAmelCase__=True , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__="gelu" , lowerCAmelCase__=False , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=32 , **lowerCAmelCase__ , ) -> Union[str, Any]: super().__init__(**lowerCamelCase_ ) a : int = image_size a : Union[str, Any] = patch_size a : int = num_channels a : int = embed_dim a : int = depths a : List[Any] = len(lowerCamelCase_ ) a : Any = num_heads a : Dict = window_size a : int = mlp_ratio a : Dict = qkv_bias a : Dict = hidden_dropout_prob a : List[Any] = attention_probs_dropout_prob a : List[str] = drop_path_rate a : Tuple = hidden_act a : Union[str, Any] = use_absolute_embeddings a : int = layer_norm_eps a : List[str] = initializer_range a : Tuple = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model a : Union[str, Any] = int(embed_dim * 2 ** (len(lowerCamelCase_ ) - 1) ) a : str = (0, 0, 0, 0)
718
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
0
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __UpperCamelCase ( _UpperCAmelCase ): lowerCamelCase : Any =['image_processor', 'tokenizer'] lowerCamelCase : Optional[int] ='AutoImageProcessor' lowerCamelCase : Any ='AutoTokenizer' def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Dict: a : List[str] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __UpperCamelCase , ) a : Dict = kwargs.pop("feature_extractor" ) a : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__UpperCamelCase , __UpperCamelCase ) a : Tuple = self.image_processor a : Optional[Any] = False def __call__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: if self._in_target_context_manager: return self.current_processor(*__UpperCamelCase , **__UpperCamelCase ) a : Optional[int] = kwargs.pop("images" , __UpperCamelCase ) a : str = kwargs.pop("text" , __UpperCamelCase ) if len(__UpperCamelCase ) > 0: a : Optional[int] = args[0] a : List[str] = args[1:] if images is None and text is None: raise ValueError("You need to specify either an `images` or `text` input to process." ) if images is not None: a : Tuple = self.image_processor(__UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) if text is not None: a : Optional[Any] = self.tokenizer(__UpperCamelCase , **__UpperCamelCase ) if text is None: return inputs elif images is None: return encodings else: a : List[Any] = encodings["input_ids"] return inputs def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Tuple: return self.tokenizer.batch_decode(*__UpperCamelCase , **__UpperCamelCase ) def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: return self.tokenizer.decode(*__UpperCamelCase , **__UpperCamelCase ) @contextmanager def __a ( self ) -> Union[str, Any]: warnings.warn( "`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your " "labels by using the argument `text` of the regular `__call__` method (either in the same call as " "your images inputs, or in a separate call." ) a : int = True a : List[str] = self.tokenizer yield a : str = self.image_processor a : Any = False def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=None ) -> List[str]: if added_vocab is None: a : Tuple = self.tokenizer.get_added_vocab() a : int = {} while tokens: a : Optional[Any] = re.search(R"<s_(.*?)>" , __UpperCamelCase , re.IGNORECASE ) if start_token is None: break a : Union[str, Any] = start_token.group(1 ) a : Optional[int] = re.search(Rf"""</s_{key}>""" , __UpperCamelCase , re.IGNORECASE ) a : Dict = start_token.group() if end_token is None: a : Tuple = tokens.replace(__UpperCamelCase , "" ) else: a : List[Any] = end_token.group() a : List[Any] = re.escape(__UpperCamelCase ) a : str = re.escape(__UpperCamelCase ) a : str = re.search(f"""{start_token_escaped}(.*?){end_token_escaped}""" , __UpperCamelCase , re.IGNORECASE ) if content is not None: a : int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node a : List[str] = self.tokenajson(__UpperCamelCase , is_inner_value=__UpperCamelCase , added_vocab=__UpperCamelCase ) if value: if len(__UpperCamelCase ) == 1: a : Union[str, Any] = value[0] a : List[Any] = value else: # leaf nodes a : Optional[int] = [] for leaf in content.split(R"<sep/>" ): a : List[str] = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": a : Tuple = leaf[1:-2] # for categorical special tokens output[key].append(__UpperCamelCase ) if len(output[key] ) == 1: a : Dict = output[key][0] a : List[str] = tokens[tokens.find(__UpperCamelCase ) + len(__UpperCamelCase ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=__UpperCamelCase , added_vocab=__UpperCamelCase ) if len(__UpperCamelCase ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def __a ( self ) -> Optional[int]: warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __UpperCamelCase , ) return self.image_processor_class @property def __a ( self ) -> Dict: warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __UpperCamelCase , ) return self.image_processor
719
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
0
from typing import TYPE_CHECKING import torch from ..models.auto import AutoModelForVisualQuestionAnswering, AutoProcessor from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __UpperCamelCase ( _a ): lowerCamelCase : Tuple ="""dandelin/vilt-b32-finetuned-vqa""" lowerCamelCase : str =( """This is a tool that answers a question about an image. It takes an input named `image` which should be the """ """image containing the information, as well as a `question` which should be the question in English. It """ """returns a text that is the answer to the question.""" ) lowerCamelCase : Optional[Any] ="""image_qa""" lowerCamelCase : Dict =AutoProcessor lowerCamelCase : int =AutoModelForVisualQuestionAnswering lowerCamelCase : str =["""image""", """text"""] lowerCamelCase : List[Any] =["""text"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[Any]: requires_backends(self , ["vision"] ) super().__init__(*snake_case_ , **snake_case_ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: return self.pre_processor(snake_case_ , snake_case_ , return_tensors="pt" ) def __a ( self , lowerCAmelCase__ ) -> Dict: with torch.no_grad(): return self.model(**snake_case_ ).logits def __a ( self , lowerCAmelCase__ ) -> List[Any]: a : List[Any] = outputs.argmax(-1 ).item() return self.model.config.idalabel[idx]
720
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
0
"""simple docstring""" import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' a : int = 10 a : Any = datasets.Features( { "tokens": datasets.Sequence(datasets.Value("string" ) ), "labels": datasets.Sequence(datasets.ClassLabel(names=["negative", "positive"] ) ), "answers": datasets.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), "id": datasets.Value("int64" ), } ) a : Tuple = datasets.Dataset.from_dict( { "tokens": [["foo"] * 5] * n, "labels": [[1] * 5] * n, "answers": [{"answer_start": [97], "text": ["1976"]}] * 10, "id": list(range(__UpperCamelCase ) ), } , features=__UpperCamelCase , ) return dataset @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Dict ) ->List[str]: '''simple docstring''' a : str = str(tmp_path_factory.mktemp("data" ) / "file.arrow" ) dataset.map(cache_file_name=__UpperCamelCase ) return filename # FILE_CONTENT + files _A : List[Any] = '''\ Text data. Second line of data.''' @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->Dict: '''simple docstring''' a : Union[str, Any] = tmp_path_factory.mktemp("data" ) / """file.txt""" a : Optional[Any] = FILE_CONTENT with open(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase ) return filename @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Optional[Any]: '''simple docstring''' import bza a : Dict = tmp_path_factory.mktemp("data" ) / """file.txt.bz2""" a : Any = bytes(__UpperCamelCase , "utf-8" ) with bza.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] ) ->Any: '''simple docstring''' import gzip a : Tuple = str(tmp_path_factory.mktemp("data" ) / "file.txt.gz" ) a : List[Any] = bytes(__UpperCamelCase , "utf-8" ) with gzip.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->str: '''simple docstring''' if datasets.config.LZ4_AVAILABLE: import lza.frame a : Any = tmp_path_factory.mktemp("data" ) / """file.txt.lz4""" a : Optional[Any] = bytes(__UpperCamelCase , "utf-8" ) with lza.frame.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Optional[int] ) ->Dict: '''simple docstring''' if datasets.config.PY7ZR_AVAILABLE: import pyazr a : Any = tmp_path_factory.mktemp("data" ) / """file.txt.7z""" with pyazr.SevenZipFile(__UpperCamelCase , "w" ) as archive: archive.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Tuple ) ->str: '''simple docstring''' import tarfile a : Tuple = tmp_path_factory.mktemp("data" ) / """file.txt.tar""" with tarfile.TarFile(__UpperCamelCase , "w" ) as f: f.add(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Any: '''simple docstring''' import lzma a : str = tmp_path_factory.mktemp("data" ) / """file.txt.xz""" a : str = bytes(__UpperCamelCase , "utf-8" ) with lzma.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Dict ) ->List[Any]: '''simple docstring''' import zipfile a : Optional[int] = tmp_path_factory.mktemp("data" ) / """file.txt.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Dict: '''simple docstring''' if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd a : Any = tmp_path_factory.mktemp("data" ) / """file.txt.zst""" a : Tuple = bytes(__UpperCamelCase , "utf-8" ) with zstd.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] ) ->Any: '''simple docstring''' a : Dict = tmp_path_factory.mktemp("data" ) / """file.xml""" a : List[str] = textwrap.dedent( "\\n <?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n <tmx version=\"1.4\">\n <header segtype=\"sentence\" srclang=\"ca\" />\n <body>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>" ) with open(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase ) return filename _A : List[str] = [ {'''col_1''': '''0''', '''col_2''': 0, '''col_3''': 0.0}, {'''col_1''': '''1''', '''col_2''': 1, '''col_3''': 1.0}, {'''col_1''': '''2''', '''col_2''': 2, '''col_3''': 2.0}, {'''col_1''': '''3''', '''col_2''': 3, '''col_3''': 3.0}, ] _A : Tuple = [ {'''col_1''': '''4''', '''col_2''': 4, '''col_3''': 4.0}, {'''col_1''': '''5''', '''col_2''': 5, '''col_3''': 5.0}, ] _A : int = { '''col_1''': ['''0''', '''1''', '''2''', '''3'''], '''col_2''': [0, 1, 2, 3], '''col_3''': [0.0, 1.0, 2.0, 3.0], } _A : int = [ {'''col_3''': 0.0, '''col_1''': '''0''', '''col_2''': 0}, {'''col_3''': 1.0, '''col_1''': '''1''', '''col_2''': 1}, ] _A : Any = [ {'''col_1''': '''s0''', '''col_2''': 0, '''col_3''': 0.0}, {'''col_1''': '''s1''', '''col_2''': 1, '''col_3''': 1.0}, {'''col_1''': '''s2''', '''col_2''': 2, '''col_3''': 2.0}, {'''col_1''': '''s3''', '''col_2''': 3, '''col_3''': 3.0}, ] @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( ) ->Optional[int]: '''simple docstring''' return DATA_DICT_OF_LISTS @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' a : Optional[int] = datasets.Dataset.from_dict(__UpperCamelCase ) a : List[str] = str(tmp_path_factory.mktemp("data" ) / "dataset.arrow" ) dataset.map(cache_file_name=__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' a : str = str(tmp_path_factory.mktemp("data" ) / "dataset.sqlite" ) with contextlib.closing(sqlitea.connect(__UpperCamelCase ) ) as con: a : Tuple = con.cursor() cur.execute("CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)" ) for item in DATA: cur.execute("INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)" , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->Dict: '''simple docstring''' a : List[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.csv" ) with open(__UpperCamelCase , "w" , newline="" ) as f: a : Optional[Any] = csv.DictWriter(__UpperCamelCase , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Any: '''simple docstring''' a : Union[str, Any] = str(tmp_path_factory.mktemp("data" ) / "dataset2.csv" ) with open(__UpperCamelCase , "w" , newline="" ) as f: a : str = csv.DictWriter(__UpperCamelCase , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : int ) ->Any: '''simple docstring''' import bza a : Tuple = tmp_path_factory.mktemp("data" ) / """dataset.csv.bz2""" with open(__UpperCamelCase , "rb" ) as f: a : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(__UpperCamelCase , "wb" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : int , _lowercase : str ) ->Union[str, Any]: '''simple docstring''' a : int = tmp_path_factory.mktemp("data" ) / """dataset.csv.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : List[str] , _lowercase : Optional[Any] ) ->Dict: '''simple docstring''' a : Any = tmp_path_factory.mktemp("data" ) / """dataset.csv.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(csv_path.replace(".csv" , ".CSV" ) ) ) f.write(__UpperCamelCase , arcname=os.path.basename(csva_path.replace(".csv" , ".CSV" ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Optional[Any] , _lowercase : Optional[int] ) ->List[Any]: '''simple docstring''' a : int = tmp_path_factory.mktemp("data" ) / """dataset_with_dir.csv.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->Tuple: '''simple docstring''' a : Any = str(tmp_path_factory.mktemp("data" ) / "dataset.parquet" ) a : Any = pa.schema( { "col_1": pa.string(), "col_2": pa.intaa(), "col_3": pa.floataa(), } ) with open(__UpperCamelCase , "wb" ) as f: a : Optional[int] = pq.ParquetWriter(__UpperCamelCase , schema=__UpperCamelCase ) a : Optional[int] = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(__UpperCamelCase ) )] for k in DATA[0]} , schema=__UpperCamelCase ) writer.write_table(__UpperCamelCase ) writer.close() return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Union[str, Any]: '''simple docstring''' a : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) a : Any = {"""data""": DATA} with open(__UpperCamelCase , "w" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->str: '''simple docstring''' a : Tuple = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) a : List[Any] = {"""data""": DATA_DICT_OF_LISTS} with open(__UpperCamelCase , "w" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->List[str]: '''simple docstring''' a : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl" ) with open(__UpperCamelCase , "w" ) as f: for item in DATA: f.write(json.dumps(__UpperCamelCase ) + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : int = str(tmp_path_factory.mktemp("data" ) / "dataset2.jsonl" ) with open(__UpperCamelCase , "w" ) as f: for item in DATA: f.write(json.dumps(__UpperCamelCase ) + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Optional[int]: '''simple docstring''' a : str = str(tmp_path_factory.mktemp("data" ) / "dataset_312.jsonl" ) with open(__UpperCamelCase , "w" ) as f: for item in DATA_312: f.write(json.dumps(__UpperCamelCase ) + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->Optional[Any]: '''simple docstring''' a : Optional[Any] = str(tmp_path_factory.mktemp("data" ) / "dataset-str.jsonl" ) with open(__UpperCamelCase , "w" ) as f: for item in DATA_STR: f.write(json.dumps(__UpperCamelCase ) + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Union[str, Any] ) ->List[str]: '''simple docstring''' import gzip a : List[str] = str(tmp_path_factory.mktemp("data" ) / "dataset.txt.gz" ) with open(__UpperCamelCase , "rb" ) as orig_file: with gzip.open(__UpperCamelCase , "wb" ) as zipped_file: zipped_file.writelines(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : List[str] ) ->Optional[int]: '''simple docstring''' import gzip a : Dict = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl.gz" ) with open(__UpperCamelCase , "rb" ) as orig_file: with gzip.open(__UpperCamelCase , "wb" ) as zipped_file: zipped_file.writelines(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : Dict , _lowercase : Union[str, Any] ) ->List[str]: '''simple docstring''' a : List[str] = tmp_path_factory.mktemp("data" ) / """dataset.jsonl.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Tuple , _lowercase : Optional[Any] , _lowercase : str ) ->str: '''simple docstring''' a : int = tmp_path_factory.mktemp("data" ) / """dataset_nested.jsonl.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.join("nested" , os.path.basename(__UpperCamelCase ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : List[Any] , _lowercase : Dict ) ->int: '''simple docstring''' a : Union[str, Any] = tmp_path_factory.mktemp("data" ) / """dataset_with_dir.jsonl.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Tuple , _lowercase : Tuple ) ->List[str]: '''simple docstring''' a : Tuple = tmp_path_factory.mktemp("data" ) / """dataset.jsonl.tar""" with tarfile.TarFile(__UpperCamelCase , "w" ) as f: f.add(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) f.add(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Any , _lowercase : List[Any] , _lowercase : int ) ->List[str]: '''simple docstring''' a : Union[str, Any] = tmp_path_factory.mktemp("data" ) / """dataset_nested.jsonl.tar""" with tarfile.TarFile(__UpperCamelCase , "w" ) as f: f.add(__UpperCamelCase , arcname=os.path.join("nested" , os.path.basename(__UpperCamelCase ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->Dict: '''simple docstring''' a : str = ["""0""", """1""", """2""", """3"""] a : Dict = str(tmp_path_factory.mktemp("data" ) / "dataset.txt" ) with open(__UpperCamelCase , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Union[str, Any]: '''simple docstring''' a : int = ["""0""", """1""", """2""", """3"""] a : int = str(tmp_path_factory.mktemp("data" ) / "dataset2.txt" ) with open(__UpperCamelCase , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->List[Any]: '''simple docstring''' a : List[Any] = ["""0""", """1""", """2""", """3"""] a : str = tmp_path_factory.mktemp("data" ) / """dataset.abc""" with open(__UpperCamelCase , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : List[Any] , _lowercase : Dict ) ->List[str]: '''simple docstring''' a : Union[str, Any] = tmp_path_factory.mktemp("data" ) / """dataset.text.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Any , _lowercase : Optional[Any] ) ->Optional[Any]: '''simple docstring''' a : Dict = tmp_path_factory.mktemp("data" ) / """dataset_with_dir.text.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) f.write(__UpperCamelCase , arcname=os.path.join("main_dir" , os.path.basename(__UpperCamelCase ) ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : Optional[Any] , _lowercase : Optional[int] ) ->Tuple: '''simple docstring''' a : Optional[Any] = tmp_path_factory.mktemp("data" ) / """dataset.ext.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename("unsupported.ext" ) ) f.write(__UpperCamelCase , arcname=os.path.basename("unsupported_2.ext" ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->Optional[int]: '''simple docstring''' a : List[Any] = """\n""".join(["First", "Second\u2029with Unicode new line", "Third"] ) a : List[str] = str(tmp_path_factory.mktemp("data" ) / "dataset_with_unicode_new_lines.txt" ) with open(__UpperCamelCase , "w" , encoding="utf-8" ) as f: f.write(__UpperCamelCase ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( ) ->Optional[Any]: '''simple docstring''' return os.path.join("tests" , "features" , "data" , "test_image_rgb.jpg" ) @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( ) ->Optional[int]: '''simple docstring''' return os.path.join("tests" , "features" , "data" , "test_audio_44100.wav" ) @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : List[Any] ) ->Any: '''simple docstring''' a : Any = tmp_path_factory.mktemp("data" ) / """dataset.img.zip""" with zipfile.ZipFile(__UpperCamelCase , "w" ) as f: f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ) ) f.write(__UpperCamelCase , arcname=os.path.basename(__UpperCamelCase ).replace(".jpg" , "2.jpg" ) ) return path @pytest.fixture(scope="session" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : str = tmp_path_factory.mktemp("data_dir" ) (data_dir / "subdir").mkdir() with open(data_dir / "subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / "subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden file with open(data_dir / "subdir" / ".test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / ".subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / ".subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) return data_dir
721
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer a : str = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} a : Tuple = { '''vocab_file''': { '''squeezebert/squeezebert-uncased''': ( '''https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt''' ), '''squeezebert/squeezebert-mnli''': '''https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt''', '''squeezebert/squeezebert-mnli-headless''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''squeezebert/squeezebert-uncased''': ( '''https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json''' ), '''squeezebert/squeezebert-mnli''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json''' ), '''squeezebert/squeezebert-mnli-headless''': ( '''https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json''' ), }, } a : int = { '''squeezebert/squeezebert-uncased''': 512, '''squeezebert/squeezebert-mnli''': 512, '''squeezebert/squeezebert-mnli-headless''': 512, } a : List[Any] = { '''squeezebert/squeezebert-uncased''': {'''do_lower_case''': True}, '''squeezebert/squeezebert-mnli''': {'''do_lower_case''': True}, '''squeezebert/squeezebert-mnli-headless''': {'''do_lower_case''': True}, } class __UpperCamelCase ( _lowercase ): lowerCamelCase : List[Any] =VOCAB_FILES_NAMES lowerCamelCase : Tuple =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : str =PRETRAINED_INIT_CONFIGURATION lowerCamelCase : Tuple =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[str] =SqueezeBertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__="[UNK]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="[PAD]" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> str: super().__init__( A_ , tokenizer_file=A_ , do_lower_case=A_ , unk_token=A_ , sep_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , tokenize_chinese_chars=A_ , strip_accents=A_ , **A_ , ) a : List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , A_ ) != do_lower_case or normalizer_state.get("strip_accents" , A_ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , A_ ) != tokenize_chinese_chars ): a : Tuple = getattr(A_ , normalizer_state.pop("type" ) ) a : Optional[int] = do_lower_case a : Tuple = strip_accents a : List[Any] = tokenize_chinese_chars a : List[Any] = normalizer_class(**A_ ) a : str = do_lower_case def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> Any: a : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Tuple = [self.sep_token_id] a : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: a : List[Any] = self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ )
700
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
0
"""simple docstring""" import gc import threading import time import psutil import torch class __UpperCamelCase : def __init__( self ) -> List[str]: a : List[str] = psutil.Process() a : Dict = False def __a ( self ) -> Any: a : str = -1 while True: a : List[Any] = max(self.process.memory_info().rss , self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def __a ( self ) -> str: a : Dict = True a : Optional[Any] = threading.Thread(target=self.peak_monitor ) a : Optional[int] = True self.thread.start() def __a ( self ) -> Any: a : Optional[int] = False self.thread.join() return self.cpu_memory_peak a : str = PeakCPUMemory() def _SCREAMING_SNAKE_CASE ( ) ->List[Any]: '''simple docstring''' a : str = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem a : Any = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): a : List[str] = torch.cuda.memory_allocated(__SCREAMING_SNAKE_CASE ) torch.cuda.reset_peak_memory_stats() return measures def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->int: '''simple docstring''' a : Optional[Any] = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem a : Tuple = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 a : Optional[Any] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): a : List[Any] = (torch.cuda.memory_allocated(__SCREAMING_SNAKE_CASE ) - start_measures[str(__SCREAMING_SNAKE_CASE )]) / 2**20 a : List[Any] = (torch.cuda.max_memory_allocated(__SCREAMING_SNAKE_CASE ) - start_measures[str(__SCREAMING_SNAKE_CASE )]) / 2**20 return measures def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : int ) ->Optional[int]: '''simple docstring''' print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__SCREAMING_SNAKE_CASE )]:.2f}MiB""" ) a : List[str] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
701
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
0
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py a : Any = 'src/diffusers' a : Tuple = '.' # This is to make sure the diffusers module imported is the one in the repo. a : Optional[Any] = importlib.util.spec_from_file_location( 'diffusers', os.path.join(DIFFUSERS_PATH, '__init__.py'), submodule_search_locations=[DIFFUSERS_PATH], ) a : Optional[int] = spec.loader.load_module() def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Tuple ) ->Dict: '''simple docstring''' return line.startswith(_lowercase ) or len(_lowercase ) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , _lowercase ) is not None def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Any: '''simple docstring''' a : Tuple = object_name.split("." ) a : Optional[Any] = 0 # First let's find the module where our object lives. a : int = parts[i] while i < len(_lowercase ) and not os.path.isfile(os.path.join(_lowercase , F"""{module}.py""" ) ): i += 1 if i < len(_lowercase ): a : List[str] = os.path.join(_lowercase , parts[i] ) if i >= len(_lowercase ): raise ValueError(F"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(_lowercase , F"""{module}.py""" ) , "r" , encoding="utf-8" , newline="\n" ) as f: a : Dict = f.readlines() # Now let's find the class / func in the code! a : Dict = "" a : Optional[int] = 0 for name in parts[i + 1 :]: while ( line_index < len(_lowercase ) and re.search(RF"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(_lowercase ): raise ValueError(F""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). a : List[Any] = line_index while line_index < len(_lowercase ) and _should_continue(lines[line_index] , _lowercase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a : int = lines[start_index:line_index] return "".join(_lowercase ) a : Optional[int] = re.compile(R'^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)') a : List[str] = re.compile(R'^\s*(\S+)->(\S+)(\s+.*|$)') a : Any = re.compile(R'<FILL\s+[^>]*>') def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Dict: '''simple docstring''' a : List[str] = code.split("\n" ) a : Optional[Any] = 0 while idx < len(_lowercase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(_lowercase ): return re.search(R"^(\s*)\S" , lines[idx] ).groups()[0] return "" def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Any: '''simple docstring''' a : Optional[Any] = len(get_indent(_lowercase ) ) > 0 if has_indent: a : Optional[Any] = F"""class Bla:\n{code}""" a : List[str] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=_lowercase ) a : List[Any] = black.format_str(_lowercase , mode=_lowercase ) a, a : List[Any] = style_docstrings_in_code(_lowercase ) return result[len("class Bla:\n" ) :] if has_indent else result def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int=False ) ->Optional[int]: '''simple docstring''' with open(_lowercase , "r" , encoding="utf-8" , newline="\n" ) as f: a : Dict = f.readlines() a : Optional[Any] = [] a : Optional[int] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(_lowercase ): a : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. a, a, a : int = search.groups() a : str = find_code_in_diffusers(_lowercase ) a : Optional[int] = get_indent(_lowercase ) a : Any = line_index + 1 if indent == theoretical_indent else line_index + 2 a : str = theoretical_indent a : List[str] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. a : Any = True while line_index < len(_lowercase ) and should_continue: line_index += 1 if line_index >= len(_lowercase ): break a : List[str] = lines[line_index] a : Optional[int] = _should_continue(_lowercase , _lowercase ) and re.search(F"""^{indent}# End copy""" , _lowercase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a : str = lines[start_index:line_index] a : int = "".join(_lowercase ) # Remove any nested `Copied from` comments to avoid circular copies a : Optional[int] = [line for line in theoretical_code.split("\n" ) if _re_copy_warning.search(_lowercase ) is None] a : Dict = "\n".join(_lowercase ) # Before comparing, use the `replace_pattern` on the original code. if len(_lowercase ) > 0: a : int = replace_pattern.replace("with" , "" ).split("," ) a : List[Any] = [_re_replace_pattern.search(_lowercase ) for p in patterns] for pattern in patterns: if pattern is None: continue a, a, a : Dict = pattern.groups() a : Any = re.sub(_lowercase , _lowercase , _lowercase ) if option.strip() == "all-casing": a : int = re.sub(obja.lower() , obja.lower() , _lowercase ) a : Tuple = re.sub(obja.upper() , obja.upper() , _lowercase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line a : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) a : List[str] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: a : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] a : Tuple = start_index + 1 if overwrite and len(_lowercase ) > 0: # Warn the user a file has been modified. print(F"""Detected changes, rewriting {filename}.""" ) with open(_lowercase , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(_lowercase ) return diffs def _SCREAMING_SNAKE_CASE ( _lowercase : bool = False ) ->int: '''simple docstring''' a : Optional[Any] = glob.glob(os.path.join(_lowercase , "**/*.py" ) , recursive=_lowercase ) a : int = [] for filename in all_files: a : List[Any] = is_copy_consistent(_lowercase , _lowercase ) diffs += [F"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(_lowercase ) > 0: a : Tuple = "\n".join(_lowercase ) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them." ) if __name__ == "__main__": a : Any = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') a : Tuple = parser.parse_args() check_copies(args.fix_and_overwrite)
702
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
0
"""simple docstring""" import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 a : Union[str, Any] = get_tests_dir('''fixtures/dummy-config.json''') class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Union[str, Any]: a : str = 0 def __a ( self ) -> Any: self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("transformers.models.auto" ) ) def __a ( self ) -> Union[str, Any]: a : str = AutoConfig.from_pretrained("bert-base-uncased" ) self.assertIsInstance(__A , __A ) def __a ( self ) -> str: a : Optional[int] = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def __a ( self ) -> List[str]: a : Any = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def __a ( self ) -> Optional[Any]: a : Tuple = AutoConfig.for_model("roberta" ) self.assertIsInstance(__A , __A ) def __a ( self ) -> Any: with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. a : Any = os.path.join(__A , "fake-roberta" ) os.makedirs(__A , exist_ok=__A ) with open(os.path.join(__A , "config.json" ) , "w" ) as f: f.write(json.dumps({} ) ) a : List[Any] = AutoConfig.from_pretrained(__A ) self.assertEqual(type(__A ) , __A ) def __a ( self ) -> Dict: try: AutoConfig.register("custom" , __A ) # Wrong model type will raise an error with self.assertRaises(__A ): AutoConfig.register("model" , __A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoConfig.register("bert" , __A ) # Now that the config is registered, it can be used as any other config with the auto-API a : str = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__A ) a : List[str] = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def __a ( self ) -> List[Any]: with self.assertRaisesRegex( __A , "bert-base is not a local folder and is not a valid model identifier" ): a : Any = AutoConfig.from_pretrained("bert-base" ) def __a ( self ) -> str: with self.assertRaisesRegex( __A , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): a : Optional[Any] = AutoConfig.from_pretrained(__A , revision="aaaaaa" ) def __a ( self ) -> Dict: with self.assertRaisesRegex( __A , "hf-internal-testing/no-config-test-repo does not appear to have a file named config.json." , ): a : Any = AutoConfig.from_pretrained("hf-internal-testing/no-config-test-repo" ) def __a ( self ) -> Any: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__A ): a : Dict = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__A ): a : Union[str, Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=__A ) a : Optional[int] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__A ) a : Union[str, Any] = AutoConfig.from_pretrained(__A , trust_remote_code=__A ) self.assertEqual(reloaded_config.__class__.__name__ , "NewModelConfig" ) def __a ( self ) -> Optional[Any]: class __UpperCamelCase ( UpperCamelCase_ ): lowerCamelCase : Optional[Any] ="""new-model""" try: AutoConfig.register("new-model" , __A ) # If remote code is not set, the default is to use local a : Any = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote code is disabled, we load the local one. a : Dict = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote is enabled, we load from the Hub a : Dict = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
703
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
0
"""simple docstring""" from random import shuffle import tensorflow as tf from numpy import array def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Dict ) ->Tuple: '''simple docstring''' a : Optional[Any] = int(_lowerCamelCase ) assert noofclusters < len(_lowerCamelCase ) # Find out the dimensionality a : str = len(vectors[0] ) # Will help select random centroids from among the available vectors a : Optional[Any] = list(range(len(_lowerCamelCase ) ) ) shuffle(_lowerCamelCase ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. a : List[Any] = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION a : Any = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points a : Union[str, Any] = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(_lowerCamelCase ) ] ##These nodes will assign the centroid Variables the appropriate ##values a : List[str] = tf.placeholder("float64" , [dim] ) a : Any = [] for centroid in centroids: cent_assigns.append(tf.assign(_lowerCamelCase , _lowerCamelCase ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) a : Dict = [tf.Variable(0 ) for i in range(len(_lowerCamelCase ) )] ##These nodes will assign an assignment Variable the appropriate ##value a : str = tf.placeholder("int32" ) a : Tuple = [] for assignment in assignments: cluster_assigns.append(tf.assign(_lowerCamelCase , _lowerCamelCase ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input a : Tuple = tf.placeholder("float" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors a : List[Any] = tf.reduce_mean(_lowerCamelCase , 0 ) ##Node for computing Euclidean distances # Placeholders for input a : List[str] = tf.placeholder("float" , [dim] ) a : str = tf.placeholder("float" , [dim] ) a : Union[str, Any] = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(_lowerCamelCase , _lowerCamelCase ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input a : Union[str, Any] = tf.placeholder("float" , [noofclusters] ) a : Optional[int] = tf.argmin(_lowerCamelCase , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. a : Any = tf.initialize_all_variables() # Initialize all variables sess.run(_lowerCamelCase ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. a : Union[str, Any] = 100 for _ in range(_lowerCamelCase ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(_lowerCamelCase ) ): a : int = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. a : Optional[int] = [ sess.run(_lowerCamelCase , feed_dict={va: vect, va: sess.run(_lowerCamelCase )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input a : Tuple = sess.run( _lowerCamelCase , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(_lowerCamelCase ): # Collect all the vectors assigned to this cluster a : Union[str, Any] = [ vectors[i] for i in range(len(_lowerCamelCase ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location a : Tuple = sess.run( _lowerCamelCase , feed_dict={mean_input: array(_lowerCamelCase )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments a : Optional[Any] = sess.run(_lowerCamelCase ) a : List[Any] = sess.run(_lowerCamelCase ) return centroids, assignments
704
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
0
"""simple docstring""" import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=30 , lowerCAmelCase__=2 , lowerCAmelCase__=3 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=10 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=0.6 , lowerCAmelCase__=None , ) -> int: a : List[str] = parent a : Optional[int] = batch_size a : Any = image_size a : Tuple = patch_size a : Any = num_channels a : Dict = is_training a : str = use_labels a : List[str] = hidden_size a : Optional[Any] = num_hidden_layers a : Optional[int] = num_attention_heads a : Optional[Any] = intermediate_size a : str = hidden_act a : Union[str, Any] = hidden_dropout_prob a : Optional[int] = attention_probs_dropout_prob a : List[str] = type_sequence_label_size a : Optional[int] = initializer_range a : Optional[Any] = mask_ratio a : Optional[Any] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) a : Optional[Any] = (image_size // patch_size) ** 2 a : Any = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def __a ( self ) -> Any: a : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a : Tuple = None if self.use_labels: a : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : Any = self.get_config() return config, pixel_values, labels def __a ( self ) -> Optional[int]: return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: a : int = ViTMAEModel(config=_a ) model.to(_a ) model.eval() a : Tuple = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Union[str, Any] = ViTMAEForPreTraining(_a ) model.to(_a ) model.eval() a : Optional[Any] = model(_a ) a : Any = (self.image_size // self.patch_size) ** 2 a : str = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images a : int = 1 a : Tuple = ViTMAEForPreTraining(_a ) model.to(_a ) model.eval() a : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) a : str = model(_a ) a : Any = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def __a ( self ) -> List[Any]: a : List[str] = self.prepare_config_and_inputs() a : Dict = config_and_inputs a : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): lowerCamelCase : str =(ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowerCamelCase : List[Any] ={"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} lowerCamelCase : Optional[int] =False lowerCamelCase : Tuple =False lowerCamelCase : Dict =False lowerCamelCase : Optional[int] =False def __a ( self ) -> str: a : List[Any] = ViTMAEModelTester(self ) a : str = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __a ( self ) -> str: self.config_tester.run_common_tests() @unittest.skip(reason="ViTMAE does not use inputs_embeds" ) def __a ( self ) -> Any: pass def __a ( self ) -> Optional[Any]: a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : List[str] = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) a : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def __a ( self ) -> int: a : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : str = model_class(_a ) a : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : Union[str, Any] = [*signature.parameters.keys()] a : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _a ) def __a ( self ) -> List[Any]: a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __a ( self ) -> List[str]: a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_a ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: np.random.seed(2 ) a : str = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) a : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) a : Optional[Any] = torch.from_numpy(_a ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument a : List[Any] = pt_noise super().check_pt_tf_models(_a , _a , _a ) def __a ( self ) -> Tuple: a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : List[str] = model_class(_a ) model.to(_a ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): a : Dict = model(**self._prepare_for_class(_a , _a ) ) a : Tuple = outputs[0].cpu().numpy() a : Optional[Any] = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_a ) a : Union[str, Any] = model_class.from_pretrained(_a ) model.to(_a ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): a : Union[str, Any] = model(**self._prepare_for_class(_a , _a ) ) # Make sure we don't have nans a : Optional[int] = after_outputs[0].cpu().numpy() a : Tuple = 0 a : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_a , 1E-5 ) @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def __a ( self ) -> List[str]: pass @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def __a ( self ) -> int: pass @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def __a ( self ) -> Optional[int]: pass @unittest.skip(reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load" ) def __a ( self ) -> Optional[int]: pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> Any: pass @slow def __a ( self ) -> int: for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Optional[int] = ViTMAEModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def _SCREAMING_SNAKE_CASE ( ) ->int: '''simple docstring''' a : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def __a ( self ) -> Dict: return ViTImageProcessor.from_pretrained("facebook/vit-mae-base" ) if is_vision_available() else None @slow def __a ( self ) -> Any: np.random.seed(2 ) a : Optional[Any] = ViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base" ).to(_a ) a : Dict = self.default_image_processor a : Union[str, Any] = prepare_img() a : List[Any] = image_processor(images=_a , return_tensors="pt" ).to(_a ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) a : Optional[int] = ViTMAEConfig() a : int = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) a : Dict = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): a : Tuple = model(**_a , noise=torch.from_numpy(_a ).to(device=_a ) ) # verify the logits a : Any = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , _a ) a : int = torch.tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(_a ) , atol=1E-4 ) )
705
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
0
"""simple docstring""" from __future__ import annotations a : List[Any] = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def _SCREAMING_SNAKE_CASE ( _lowercase : list[list[int]] , _lowercase : list[int] , _lowercase : list[int] , _lowercase : int , _lowercase : list[list[int]] , ) ->List[Any]: '''simple docstring''' a : List[Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(snake_case__ ) ) ] # the reference grid a : List[str] = 1 a : Union[str, Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(snake_case__ ) ) ] # the action grid a : List[Any] = init[0] a : str = init[1] a : int = 0 a : Optional[int] = g + heuristic[x][y] # cost from starting cell to destination cell a : int = [[f, g, x, y]] a : Union[str, Any] = False # flag that is set when search is complete a : int = False # flag set if we can't find expand while not found and not resign: if len(snake_case__ ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() a : Dict = cell.pop() a : str = next_cell[2] a : List[str] = next_cell[3] a : List[Any] = next_cell[1] if x == goal[0] and y == goal[1]: a : Optional[int] = True else: for i in range(len(snake_case__ ) ): # to try out different valid actions a : Dict = x + DIRECTIONS[i][0] a : Dict = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(snake_case__ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: a : List[str] = g + cost a : str = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) a : str = 1 a : Any = i a : List[Any] = [] a : List[str] = goal[0] a : List[Any] = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: a : Any = x - DIRECTIONS[action[x][y]][0] a : List[Any] = y - DIRECTIONS[action[x][y]][1] a : Optional[int] = xa a : int = ya invpath.append([x, y] ) a : int = [] for i in range(len(snake_case__ ) ): path.append(invpath[len(snake_case__ ) - 1 - i] ) return path, action if __name__ == "__main__": a : Optional[Any] = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] a : Optional[int] = [0, 0] # all coordinates are given in format [y,x] a : Tuple = [len(grid) - 1, len(grid[0]) - 1] a : Any = 1 # the cost map which pushes the path closer to the goal a : Any = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): a : Tuple = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map a : int = 99 a , a : int = search(grid, init, goal, cost, heuristic) print('''ACTION MAP''') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
706
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
0
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Optional[int] , _lowercase : Optional[Any] ) ->float: '''simple docstring''' if days_between_payments <= 0: raise ValueError("days_between_payments must be > 0" ) if daily_interest_rate < 0: raise ValueError("daily_interest_rate must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return principal * daily_interest_rate * days_between_payments def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : int , _lowercase : List[str] , ) ->float: '''simple docstring''' if number_of_compounding_periods <= 0: raise ValueError("number_of_compounding_periods must be > 0" ) if nominal_annual_interest_rate_percentage < 0: raise ValueError("nominal_annual_interest_rate_percentage must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : int , _lowercase : Optional[int] , ) ->float: '''simple docstring''' if number_of_years <= 0: raise ValueError("number_of_years must be > 0" ) if nominal_annual_percentage_rate < 0: raise ValueError("nominal_annual_percentage_rate must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return compound_interest( __snake_case , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
707
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { "configuration_xmod": [ "XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP", "XmodConfig", "XmodOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ "XMOD_PRETRAINED_MODEL_ARCHIVE_LIST", "XmodForCausalLM", "XmodForMaskedLM", "XmodForMultipleChoice", "XmodForQuestionAnswering", "XmodForSequenceClassification", "XmodForTokenClassification", "XmodModel", "XmodPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys a : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
708
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
0
"""simple docstring""" import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a : Any = get_logger(__name__) class __UpperCamelCase ( enum.Enum ): lowerCamelCase : Union[str, Any] ="""all_checks""" lowerCamelCase : int ="""basic_checks""" lowerCamelCase : Dict ="""no_checks""" class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[dict] , _lowercase : dict , _lowercase : Union[str, Any]=None ) ->Any: '''simple docstring''' if expected_checksums is None: logger.info("Unable to verify checksums." ) return if len(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) ) if len(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) > 0: raise UnexpectedDownloadedFile(str(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) ) a : Union[str, Any] = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] a : Any = " for " + verification_name if verification_name is not None else "" if len(UpperCAmelCase__ ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" "Set `verification_mode='no_checks'` to skip checksums verification and ignore this error" ) logger.info("All the checksums matched successfully" + for_verification_name ) class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass class __UpperCamelCase ( UpperCAmelCase__ ): pass def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[dict] , _lowercase : dict ) ->Union[str, Any]: '''simple docstring''' if expected_splits is None: logger.info("Unable to verify splits sizes." ) return if len(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) > 0: raise ExpectedMoreSplits(str(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) ) if len(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) > 0: raise UnexpectedSplits(str(set(UpperCAmelCase__ ) - set(UpperCAmelCase__ ) ) ) a : str = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(UpperCAmelCase__ ) > 0: raise NonMatchingSplitsSizesError(str(UpperCAmelCase__ ) ) logger.info("All the splits matched successfully." ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : bool = True ) ->Optional[Any]: '''simple docstring''' if record_checksum: a : List[str] = shaaaa() with open(UpperCAmelCase__ , "rb" ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B"" ): m.update(UpperCAmelCase__ ) a : Optional[Any] = m.hexdigest() else: a : int = None return {"num_bytes": os.path.getsize(UpperCAmelCase__ ), "checksum": checksum} def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->Optional[Any]: '''simple docstring''' if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
709
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
0
"""simple docstring""" import os from datetime import datetime as dt from github import Github a : Optional[Any] = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''enhancement''', '''new pipeline/model''', '''new scheduler''', '''wip''', ] def _SCREAMING_SNAKE_CASE ( ) ->Optional[Any]: '''simple docstring''' a : List[str] = Github(os.environ["GITHUB_TOKEN"] ) a : Union[str, Any] = g.get_repo("huggingface/diffusers" ) a : Union[str, Any] = repo.get_issues(state="open" ) for issue in open_issues: a : Tuple = sorted(issue.get_comments() , key=lambda _lowercase : i.created_at , reverse=_snake_case ) a : Dict = comments[0] if len(_snake_case ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="closed" ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="open" ) issue.remove_from_labels("stale" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) issue.add_to_labels("stale" ) if __name__ == "__main__": main()
710
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
0
"""simple docstring""" import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging a : List[Any] = { 'cola': 2, 'mnli': 3, 'mrpc': 2, 'sst-2': 2, 'sts-b': 1, 'qqp': 2, 'qnli': 2, 'rte': 2, 'wnli': 2, } logging.set_verbosity_info() def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Any , _lowercase : str , _lowercase : List[Any]=None ) ->Union[str, Any]: '''simple docstring''' a : Optional[int] = XLNetConfig.from_json_file(_lowercase ) a : List[Any] = finetuning_task.lower() if finetuning_task is not None else '''''' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) a : int = finetuning_task a : Dict = GLUE_TASKS_NUM_LABELS[finetuning_task] a : Union[str, Any] = XLNetForSequenceClassification(_lowercase ) elif "squad" in finetuning_task: a : int = finetuning_task a : Union[str, Any] = XLNetForQuestionAnswering(_lowercase ) else: a : Dict = XLNetLMHeadModel(_lowercase ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(_lowercase , _lowercase , _lowercase ) # Save pytorch-model a : str = os.path.join(_lowercase , _lowercase ) a : Union[str, Any] = os.path.join(_lowercase , _lowercase ) print(F"""Save PyTorch model to {os.path.abspath(_lowercase )}""" ) torch.save(model.state_dict() , _lowercase ) print(F"""Save configuration file to {os.path.abspath(_lowercase )}""" ) with open(_lowercase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) a : List[Any] = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
711
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
0
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] ) ->list: '''simple docstring''' if n_term == "": return [] a : List[str] = [] for temp in range(int(__UpperCamelCase ) ): series.append(F"""1/{temp + 1}""" if series else "1" ) return series if __name__ == "__main__": a : Tuple = input('''Enter the last number (nth term) of the Harmonic Series''') print('''Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n''') print(harmonic_series(nth_term))
712
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Any = logging.get_logger(__name__) class __UpperCamelCase ( UpperCAmelCase__ ): lowerCamelCase : Tuple =['pixel_values'] def __init__( self , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = PIL.Image.BICUBIC , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = 1 / 255 , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> None: super().__init__(**lowerCAmelCase__ ) a : str = size if size is not None else {"""height""": 256, """width""": 256} a : Union[str, Any] = get_size_dict(lowerCAmelCase__ ) a : List[Any] = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} a : Union[str, Any] = get_size_dict(lowerCAmelCase__ , param_name="crop_size" ) a : List[str] = do_resize a : List[str] = size a : Optional[int] = resample a : Any = do_center_crop a : Optional[Any] = crop_size a : Any = do_rescale a : Union[str, Any] = rescale_factor a : List[str] = do_normalize a : Tuple = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN a : Union[str, Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = PIL.Image.BICUBIC , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: a : Any = get_size_dict(lowerCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}""" ) return resize( lowerCAmelCase__ , size=(size["height"], size["width"]) , resample=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: a : List[Any] = get_size_dict(lowerCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}""" ) return center_crop(lowerCAmelCase__ , size=(size["height"], size["width"]) , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Tuple: return rescale(lowerCAmelCase__ , scale=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: return normalize(lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__=None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = ChannelDimension.FIRST , **lowerCAmelCase__ , ) -> PIL.Image.Image: a : Any = do_resize if do_resize is not None else self.do_resize a : Optional[Any] = resample if resample is not None else self.resample a : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop a : Dict = do_rescale if do_rescale is not None else self.do_rescale a : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor a : Dict = do_normalize if do_normalize is not None else self.do_normalize a : Optional[Any] = image_mean if image_mean is not None else self.image_mean a : Union[str, Any] = image_std if image_std is not None else self.image_std a : Union[str, Any] = size if size is not None else self.size a : List[str] = get_size_dict(lowerCAmelCase__ ) a : List[str] = crop_size if crop_size is not None else self.crop_size a : List[Any] = get_size_dict(lowerCAmelCase__ , param_name="crop_size" ) a : int = make_list_of_images(lowerCAmelCase__ ) if not valid_images(lowerCAmelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. a : Optional[Any] = [to_numpy_array(lowerCAmelCase__ ) for image in images] if do_resize: a : List[str] = [self.resize(image=lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ ) for image in images] if do_center_crop: a : Optional[Any] = [self.center_crop(image=lowerCAmelCase__ , size=lowerCAmelCase__ ) for image in images] if do_rescale: a : Tuple = [self.rescale(image=lowerCAmelCase__ , scale=lowerCAmelCase__ ) for image in images] if do_normalize: a : str = [self.normalize(image=lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ ) for image in images] a : List[str] = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__ ) for image in images] a : List[Any] = {"""pixel_values""": images} return BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__ )
713
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : Optional[int] = logging.get_logger(__name__) a : int = { '''EleutherAI/gpt-neox-20b''': '''https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json''', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class __UpperCamelCase ( lowercase__ ): lowerCamelCase : Union[str, Any] ="""gpt_neox""" def __init__( self , lowerCAmelCase__=5_0432 , lowerCAmelCase__=6144 , lowerCAmelCase__=44 , lowerCAmelCase__=64 , lowerCAmelCase__=2_4576 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.25 , lowerCAmelCase__=1_0000 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__=2048 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=True , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> List[str]: super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase ) a : List[str] = vocab_size a : Dict = max_position_embeddings a : Any = hidden_size a : Any = num_hidden_layers a : List[str] = num_attention_heads a : Optional[int] = intermediate_size a : str = hidden_act a : List[str] = rotary_pct a : Optional[int] = rotary_emb_base a : Optional[int] = attention_dropout a : Dict = hidden_dropout a : Optional[Any] = classifier_dropout a : Dict = initializer_range a : str = layer_norm_eps a : List[Any] = use_cache a : Union[str, Any] = tie_word_embeddings a : Optional[int] = use_parallel_residual a : Optional[int] = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( "The hidden size is not divisble by the number of attention heads! Make sure to update them!" ) def __a ( self ) -> str: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __lowercase ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f"""got {self.rope_scaling}""" ) a : int = self.rope_scaling.get("type" , __lowercase ) a : List[str] = self.rope_scaling.get("factor" , __lowercase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(__lowercase , __lowercase ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}""" )
714
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() a : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : Optional[Any]=False ) ->Union[str, Any]: '''simple docstring''' a : Any = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" a : str = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Union[str, Any] , _lowercase : Any=False ) ->Dict: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: a : int = "" else: a : Dict = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) a : int = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) a : Any = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict a : str = in_proj_weight[ : config.hidden_size, : ] a : int = in_proj_bias[: config.hidden_size] a : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] a : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] a : Tuple = in_proj_weight[ -config.hidden_size :, : ] a : Any = in_proj_bias[-config.hidden_size :] def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Tuple: '''simple docstring''' a : str = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : List[Any] , _lowercase : List[Any] ) ->Any: '''simple docstring''' a : Any = dct.pop(_lowerCAmelCase ) a : Union[str, Any] = val def _SCREAMING_SNAKE_CASE ( ) ->Any: '''simple docstring''' a : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" a : str = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Any , _lowercase : List[str]=False ) ->List[str]: '''simple docstring''' a : Any = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_lowerCAmelCase , ) a : List[Any] = ViTHybridConfig(backbone_config=_lowerCAmelCase , image_size=384 , num_labels=1000 ) a : Tuple = False # load original model from timm a : int = timm.create_model(_lowerCAmelCase , pretrained=_lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys a : Optional[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_lowerCAmelCase ) a : Any = create_rename_keys(_lowerCAmelCase , _lowerCAmelCase ) for src, dest in rename_keys: rename_key(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) a : Union[str, Any] = "huggingface/label-files" a : List[str] = "imagenet-1k-id2label.json" a : int = json.load(open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="dataset" ) , "r" ) ) a : List[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} a : Tuple = idalabel a : Optional[Any] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": a : Optional[int] = ViTHybridModel(_lowerCAmelCase ).eval() else: a : Any = ViTHybridForImageClassification(_lowerCAmelCase ).eval() model.load_state_dict(_lowerCAmelCase ) # create image processor a : Dict = create_transform(**resolve_data_config({} , model=_lowerCAmelCase ) ) a : Union[str, Any] = transform.transforms a : List[Any] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } a : Tuple = ViTHybridImageProcessor( do_resize=_lowerCAmelCase , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_lowerCAmelCase , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_lowerCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) a : Tuple = prepare_img() a : Dict = transform(_lowerCAmelCase ).unsqueeze(0 ) a : Tuple = processor(_lowerCAmelCase , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_lowerCAmelCase , _lowerCAmelCase ) # verify logits with torch.no_grad(): a : Optional[Any] = model(_lowerCAmelCase ) a : str = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: a : Union[str, Any] = timm_model.forward_features(_lowerCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_lowerCAmelCase , outputs.pooler_output , atol=1E-3 ) else: a : List[str] = timm_model(_lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCAmelCase , outputs.logits , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: print(F"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(F"""ybelkada/{vit_name}""" ) processor.push_to_hub(F"""ybelkada/{vit_name}""" ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--vit_name''', default='''vit_base_r50_s16_384''', type=str, help='''Name of the hybrid ViT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to upload the model to the HuggingFace hub.''' ) a : Union[str, Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
715
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
0
"""simple docstring""" from __future__ import annotations a : List[Any] = list[tuple[int, int]] a : str = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] a : Optional[Any] = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> int: a : Union[str, Any] = pos_x a : List[str] = pos_y a : List[Any] = (pos_y, pos_x) a : List[str] = goal_x a : str = goal_y a : List[Any] = g_cost a : Any = parent a : Tuple = self.calculate_heuristic() def __a ( self ) -> float: a : Optional[Any] = abs(self.pos_x - self.goal_x ) a : List[Any] = abs(self.pos_y - self.goal_y ) return dx + dy def __lt__( self , lowerCAmelCase__ ) -> bool: return self.f_cost < other.f_cost class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Dict = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCAmelCase__ ) a : List[str] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowerCAmelCase__ ) a : Any = [self.start] a : str = [] a : str = False def __a ( self ) -> Path | None: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() a : Any = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: a : Optional[Any] = True return self.retrace_path(lowerCAmelCase__ ) self.closed_nodes.append(lowerCAmelCase__ ) a : List[str] = self.get_successors(lowerCAmelCase__ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase__ ) else: # retrieve the best current path a : List[str] = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase__ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase__ ) else: self.open_nodes.append(lowerCAmelCase__ ) if not self.reached: return [self.start.pos] return None def __a ( self , lowerCAmelCase__ ) -> list[Node]: a : str = [] for action in delta: a : Dict = parent.pos_x + action[1] a : List[Any] = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCAmelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase__ , lowerCAmelCase__ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCAmelCase__ , ) ) return successors def __a ( self , lowerCAmelCase__ ) -> Path: a : str = node a : Union[str, Any] = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) a : List[Any] = current_node.parent path.reverse() return path if __name__ == "__main__": a : Dict = (0, 0) a : Optional[int] = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) print('''------''') a : List[str] = GreedyBestFirst(init, goal) a : Any = greedy_bf.search() if path: for pos_x, pos_y in path: a : str = 2 for elem in grid: print(elem)
716
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
0
"""simple docstring""" import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->Union[str, Any]: '''simple docstring''' a : List[Any] = [] if isinstance(__lowerCAmelCase , __lowerCAmelCase ): for v in tree.values(): shapes.extend(_fetch_dims(__lowerCAmelCase ) ) elif isinstance(__lowerCAmelCase , (list, tuple) ): for t in tree: shapes.extend(_fetch_dims(__lowerCAmelCase ) ) elif isinstance(__lowerCAmelCase , torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError("Not supported" ) return shapes @torch.jit.ignore def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : Dict ) ->Tuple: '''simple docstring''' a : List[Any] = [] for d in reversed(__lowerCAmelCase ): idx.append(flat_idx % d ) a : Dict = flat_idx // d return tuple(reversed(__lowerCAmelCase ) ) @torch.jit.ignore def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : Optional[int] , _lowercase : Union[str, Any] = None , _lowercase : Optional[Any] = None , ) ->List[Any]: '''simple docstring''' def reduce_edge_list(_lowercase : Union[str, Any] ) -> None: a : List[Any] = True for i in range(len(__lowerCAmelCase ) ): a : Tuple = -1 * (i + 1) l[reversed_idx] &= tally a : str = l[reversed_idx] if start_edges is None: a : Union[str, Any] = [s == 0 for s in start] reduce_edge_list(__lowerCAmelCase ) if end_edges is None: a : List[str] = [e == (d - 1) for e, d in zip(__lowerCAmelCase , __lowerCAmelCase )] reduce_edge_list(__lowerCAmelCase ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(__lowerCAmelCase ) == 0: return [()] elif len(__lowerCAmelCase ) == 1: return [(slice(start[0] , end[0] + 1 ),)] a : List[Tuple[slice, ...]] = [] a : List[slice] = [] # Dimensions common to start and end can be selected directly for s, e in zip(__lowerCAmelCase , __lowerCAmelCase ): if s == e: path_list.append(slice(__lowerCAmelCase , s + 1 ) ) else: break a : Tuple[slice, ...] = tuple(__lowerCAmelCase ) a : Tuple = len(__lowerCAmelCase ) # start == end, and we're done if divergence_idx == len(__lowerCAmelCase ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None a : List[str] = start[divergence_idx] return tuple( path + (slice(__lowerCAmelCase , sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] , [d - 1 for d in dims[divergence_idx + 1 :]] , dims[divergence_idx + 1 :] , start_edges=start_edges[divergence_idx + 1 :] , end_edges=[True for _ in end_edges[divergence_idx + 1 :]] , ) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None a : int = end[divergence_idx] return tuple( path + (slice(__lowerCAmelCase , edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] , end[divergence_idx + 1 :] , dims[divergence_idx + 1 :] , start_edges=[True for _ in start_edges[divergence_idx + 1 :]] , end_edges=end_edges[divergence_idx + 1 :] , ) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) a : str = end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Dict , _lowercase : str , _lowercase : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' a : Any = t.shape[:no_batch_dims] a : Dict = list(_flat_idx_to_idx(__lowerCAmelCase , __lowerCAmelCase ) ) # _get_minimal_slice_set is inclusive a : Optional[int] = list(_flat_idx_to_idx(flat_end - 1 , __lowerCAmelCase ) ) # Get an ordered list of slices to perform a : Union[str, Any] = _get_minimal_slice_set( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) a : Optional[Any] = [t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : int , _lowercase : str , _lowercase : List[str] , _lowercase : Tuple = False , _lowercase : int = None , _lowercase : Tuple = False , ) ->str: '''simple docstring''' if not (len(__lowerCAmelCase ) > 0): raise ValueError("Must provide at least one input" ) a : int = [shape[:no_batch_dims] for shape in _fetch_dims(__lowerCAmelCase )] a : Optional[int] = tuple([max(__lowerCAmelCase ) for s in zip(*__lowerCAmelCase )] ) def _prep_inputs(_lowercase : str ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: a : Optional[Any] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) a : str = t.reshape(-1 , *t.shape[no_batch_dims:] ) else: a : List[str] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t a : Dict[str, Any] = tensor_tree_map(_prep_inputs , __lowerCAmelCase ) a : List[Any] = None if _out is not None: a : Optional[int] = tensor_tree_map(lambda _lowercase : t.view([-1] + list(t.shape[no_batch_dims:] ) ) , _out ) a : List[Any] = 1 for d in orig_batch_dims: flat_batch_dim *= d a : List[Any] = flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(_lowercase : Optional[Any] ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t a : Tuple = 0 a : Any = prepped_outputs for _ in range(__lowerCAmelCase ): # Chunk the input if not low_mem: a : Optional[int] = _select_chunk else: a : Tuple = partial( _chunk_slice , flat_start=__lowerCAmelCase , flat_end=min(__lowerCAmelCase , i + chunk_size ) , no_batch_dims=len(__lowerCAmelCase ) , ) a : Dict[str, Any] = tensor_tree_map(__lowerCAmelCase , __lowerCAmelCase ) # Run the layer on the chunk a : Optional[int] = layer(**__lowerCAmelCase ) # Allocate space for the output if out is None: a : Optional[Any] = tensor_tree_map(lambda _lowercase : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) , __lowerCAmelCase ) # Put the chunk in its pre-allocated space if isinstance(__lowerCAmelCase , __lowerCAmelCase ): def assign(_lowercase : Union[str, Any] , _lowercase : List[Any] ) -> None: for k, v in da.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): assign(__lowerCAmelCase , da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: a : str = da[k] assign(__lowerCAmelCase , __lowerCAmelCase ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): for xa, xa in zip(__lowerCAmelCase , __lowerCAmelCase ): if _add_into_out: xa[i : i + chunk_size] += xa else: a : Any = xa elif isinstance(__lowerCAmelCase , torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: a : int = output_chunk else: raise ValueError("Not supported" ) i += chunk_size a : Union[str, Any] = tensor_tree_map(lambda _lowercase : t.view(orig_batch_dims + t.shape[1:] ) , __lowerCAmelCase ) return out class __UpperCamelCase : def __init__( self , lowerCAmelCase__ = 512 , ) -> str: a : str = max_chunk_size a : Optional[int] = None a : Optional[tuple] = None def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: logging.info("Tuning chunk size..." ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size a : List[int] = [2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] a : List[str] = [c for c in candidates if c > min_chunk_size] a : int = [min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(lowerCAmelCase__ ) -> bool: try: with torch.no_grad(): fn(*lowerCamelCase__ , chunk_size=lowerCamelCase__ ) return True except RuntimeError: return False a : List[Any] = 0 a : Tuple = len(lowerCamelCase__ ) - 1 while i > min_viable_chunk_size_index: a : List[Any] = test_chunk_size(candidates[i] ) if not viable: a : Union[str, Any] = (min_viable_chunk_size_index + i) // 2 else: a : Optional[Any] = i a : int = (i + len(lowerCamelCase__ ) - 1) // 2 return candidates[min_viable_chunk_size_index] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> bool: a : List[str] = True for aa, aa in zip(lowerCamelCase__ , lowerCamelCase__ ): assert type(lowerCamelCase__ ) == type(lowerCamelCase__ ) if isinstance(lowerCamelCase__ , (list, tuple) ): consistent &= self._compare_arg_caches(lowerCamelCase__ , lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): a : str = [v for _, v in sorted(aa.items() , key=lambda lowerCAmelCase__ : x[0] )] a : List[Any] = [v for _, v in sorted(aa.items() , key=lambda lowerCAmelCase__ : x[0] )] consistent &= self._compare_arg_caches(lowerCamelCase__ , lowerCamelCase__ ) else: consistent &= aa == aa return consistent def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> int: a : Optional[Any] = True a : tuple = tree_map(lambda lowerCAmelCase__ : a.shape if isinstance(lowerCamelCase__ , torch.Tensor ) else a , lowerCamelCase__ , lowerCamelCase__ ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(lowerCamelCase__ ) a : Dict = self._compare_arg_caches(self.cached_arg_data , lowerCamelCase__ ) else: # Otherwise, we can reuse the precomputed value a : int = False if not consistent: a : Any = self._determine_favorable_chunk_size( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) a : str = arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
717
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
0
"""simple docstring""" from typing import List from .keymap import KEYMAP, get_character def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->List[str]: '''simple docstring''' def decorator(_lowercase : Optional[int] ): a : Tuple = getattr(lowercase__ , "handle_key" , [] ) handle += [key] setattr(lowercase__ , "handle_key" , lowercase__ ) return func return decorator def _SCREAMING_SNAKE_CASE ( *_lowercase : Dict ) ->List[str]: '''simple docstring''' def decorator(_lowercase : Optional[Any] ): a : int = getattr(lowercase__ , "handle_key" , [] ) handle += keys setattr(lowercase__ , "handle_key" , lowercase__ ) return func return decorator class __UpperCamelCase ( __a ): def __new__( cls , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Tuple = super().__new__(cls , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if not hasattr(lowerCAmelCase_ , "key_handler" ): setattr(lowerCAmelCase_ , "key_handler" , {} ) setattr(lowerCAmelCase_ , "handle_input" , KeyHandler.handle_input ) for value in attrs.values(): a : Optional[Any] = getattr(lowerCAmelCase_ , "handle_key" , [] ) for key in handled_keys: a : int = value return new_cls @staticmethod def __a ( cls ) -> Optional[int]: a : Tuple = get_character() if char != KEYMAP["undefined"]: a : Optional[Any] = ord(lowerCAmelCase_ ) a : int = cls.key_handler.get(lowerCAmelCase_ ) if handler: a : List[str] = char return handler(cls ) else: return None def _SCREAMING_SNAKE_CASE ( cls : Any ) ->Union[str, Any]: '''simple docstring''' return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
718
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
0