code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase ): _snake_case = value _snake_case = None _snake_case = None class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase ): _snake_case = tree def UpperCamelCase( self , lowerCamelCase ): if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ): yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __magic_name__ : Optional[Any] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = '''AutoTokenizer''' UpperCAmelCase__ : Tuple = ['''tokenizer'''] UpperCAmelCase__ : Union[str, Any] = { '''semantic_prompt''': 1, '''coarse_prompt''': 2, '''fine_prompt''': 2, } def __init__( self , lowerCamelCase , lowerCamelCase=None ): super().__init__(lowerCamelCase ) _snake_case = speaker_embeddings @classmethod def UpperCamelCase( cls , lowerCamelCase , lowerCamelCase="speaker_embeddings_path.json" , **lowerCamelCase ): if speaker_embeddings_dict_path is not None: _snake_case = get_file_from_repo( lowerCamelCase , lowerCamelCase , subfolder=kwargs.pop("subfolder" , lowerCamelCase ) , cache_dir=kwargs.pop("cache_dir" , lowerCamelCase ) , force_download=kwargs.pop("force_download" , lowerCamelCase ) , proxies=kwargs.pop("proxies" , lowerCamelCase ) , resume_download=kwargs.pop("resume_download" , lowerCamelCase ) , local_files_only=kwargs.pop("local_files_only" , lowerCamelCase ) , use_auth_token=kwargs.pop("use_auth_token" , lowerCamelCase ) , revision=kwargs.pop("revision" , lowerCamelCase ) , ) if speaker_embeddings_path is None: logger.warning( F'''`{os.path.join(lowerCamelCase , lowerCamelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _snake_case = None else: with open(lowerCamelCase ) as speaker_embeddings_json: _snake_case = json.load(lowerCamelCase ) else: _snake_case = None _snake_case = AutoTokenizer.from_pretrained(lowerCamelCase , **lowerCamelCase ) return cls(tokenizer=lowerCamelCase , speaker_embeddings=lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase="speaker_embeddings_path.json" , lowerCamelCase="speaker_embeddings" , lowerCamelCase = False , **lowerCamelCase , ): if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase , lowerCamelCase , "v2" ) , exist_ok=lowerCamelCase ) _snake_case = {} _snake_case = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _snake_case = self._load_voice_preset(lowerCamelCase ) _snake_case = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , lowerCamelCase , F'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=lowerCamelCase , ) _snake_case = os.path.join(lowerCamelCase , F'''{prompt_key}_{key}.npy''' ) _snake_case = tmp_dict with open(os.path.join(lowerCamelCase , lowerCamelCase ) , "w" ) as fp: json.dump(lowerCamelCase , lowerCamelCase ) super().save_pretrained(lowerCamelCase , lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase = None , **lowerCamelCase ): _snake_case = self.speaker_embeddings[voice_preset] _snake_case = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _snake_case = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , lowerCamelCase ) , cache_dir=kwargs.pop("cache_dir" , lowerCamelCase ) , force_download=kwargs.pop("force_download" , lowerCamelCase ) , proxies=kwargs.pop("proxies" , lowerCamelCase ) , resume_download=kwargs.pop("resume_download" , lowerCamelCase ) , local_files_only=kwargs.pop("local_files_only" , lowerCamelCase ) , use_auth_token=kwargs.pop("use_auth_token" , lowerCamelCase ) , revision=kwargs.pop("revision" , lowerCamelCase ) , ) if path is None: raise ValueError( F'''`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _snake_case = np.load(lowerCamelCase ) return voice_preset_dict def UpperCamelCase( self , lowerCamelCase = None ): for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="pt" , lowerCamelCase=256 , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=False , **lowerCamelCase , ): if voice_preset is not None and not isinstance(lowerCamelCase , lowerCamelCase ): if ( isinstance(lowerCamelCase , lowerCamelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _snake_case = self._load_voice_preset(lowerCamelCase ) else: if isinstance(lowerCamelCase , lowerCamelCase ) and not voice_preset.endswith(".npz" ): _snake_case = voice_preset + ".npz" _snake_case = np.load(lowerCamelCase ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase , **lowerCamelCase ) _snake_case = BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase ) _snake_case = self.tokenizer( lowerCamelCase , return_tensors=lowerCamelCase , padding="max_length" , max_length=lowerCamelCase , return_attention_mask=lowerCamelCase , return_token_type_ids=lowerCamelCase , add_special_tokens=lowerCamelCase , **lowerCamelCase , ) if voice_preset is not None: _snake_case = voice_preset return encoded_text
672
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __magic_name__ : Optional[int] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"shortest_edge": 256} _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) _snake_case = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
1
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = factor * value _snake_case = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
672
'''simple docstring''' import baseaa def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaadecode(SCREAMING_SNAKE_CASE__ ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' from __future__ import annotations import requests __magic_name__ : List[str] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = "new" , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' _snake_case = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(SCREAMING_SNAKE_CASE__ ) - valid_terms ) ): _snake_case = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(SCREAMING_SNAKE_CASE__ ) _snake_case = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"User-agent": "A random string"} , ) if response.status_code == 4_29: raise requests.HTTPError _snake_case = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(SCREAMING_SNAKE_CASE__ )} _snake_case = {} for id_ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
672
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
672
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
1
'''simple docstring''' from __future__ import annotations def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2 _snake_case = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(SCREAMING_SNAKE_CASE__ ) if n > 1: factors.append(SCREAMING_SNAKE_CASE__ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
1
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
672
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = text_path elif issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [text_path] _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=("train",) ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for split in splits: _snake_case = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader({"train": text_path} , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader({"train": text_path} , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if split: _snake_case = {split: text_path} else: _snake_case = "train" _snake_case = {"train": text_path, "test": text_path} _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
672
1
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = 10_00 ): '''simple docstring''' _snake_case = 1 _snake_case = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE__ , digit + 1 ): _snake_case = [] _snake_case = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE__ ): _snake_case = len(SCREAMING_SNAKE_CASE__ ) _snake_case = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE__ ) _snake_case = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ : Any = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __magic_name__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=3 , lowerCamelCase=32 , lowerCamelCase=3 , lowerCamelCase=10 , lowerCamelCase=[10, 20, 30, 40] , lowerCamelCase=[1, 1, 2, 1] , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="relu" , lowerCamelCase=3 , lowerCamelCase=None , ): _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = embeddings_size _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = hidden_act _snake_case = num_labels _snake_case = scope _snake_case = len(lowerCamelCase ) def UpperCamelCase( self ): _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.num_labels ) _snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase( self ): return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = TFRegNetModel(config=lowerCamelCase ) _snake_case = model(lowerCamelCase , training=lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = self.num_labels _snake_case = TFRegNetForImageClassification(lowerCamelCase ) _snake_case = model(lowerCamelCase , labels=lowerCamelCase , training=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase( self ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Any = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () UpperCAmelCase__ : Union[str, Any] = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) UpperCAmelCase__ : int = False UpperCAmelCase__ : Any = False UpperCAmelCase__ : Union[str, Any] = False UpperCAmelCase__ : int = False UpperCAmelCase__ : Optional[int] = False def UpperCamelCase( self ): _snake_case = TFRegNetModelTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def UpperCamelCase( self ): return @unittest.skip(reason="RegNet does not use inputs_embeds" ) def UpperCamelCase( self ): pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("GPU" ) ) == 0 , reason="TF does not support backprop for grouped convolutions on CPU." , ) @slow def UpperCamelCase( self ): super().test_keras_fit() @unittest.skip(reason="RegNet does not support input and output embeddings" ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(lowerCamelCase ) _snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): def check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = model_class(lowerCamelCase ) _snake_case = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) , training=lowerCamelCase ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: _snake_case = layer_type _snake_case = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase={} ): _snake_case = model(lowerCamelCase , return_dict=lowerCamelCase , **lowerCamelCase ) _snake_case = model(lowerCamelCase , return_dict=lowerCamelCase , **lowerCamelCase ).to_tuple() def recursive_check(lowerCamelCase , lowerCamelCase ): if isinstance(lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowerCamelCase , lowerCamelCase ): recursive_check(lowerCamelCase , lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(lowerCamelCase , lowerCamelCase ) ) , msg=( "Tuple and dict output are not equal. Difference:" F''' {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}''' ) , ) recursive_check(lowerCamelCase , lowerCamelCase ) for model_class in self.all_model_classes: _snake_case = model_class(lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) check_equivalence(lowerCamelCase , lowerCamelCase , lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) check_equivalence(lowerCamelCase , lowerCamelCase , lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) check_equivalence(lowerCamelCase , lowerCamelCase , lowerCamelCase , {"output_hidden_states": True} ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) check_equivalence(lowerCamelCase , lowerCamelCase , lowerCamelCase , {"output_hidden_states": True} ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def UpperCamelCase( self ): for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = TFRegNetModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' _snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase( self ): return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase( self ): _snake_case = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=lowerCamelCase , return_tensors="tf" ) # forward pass _snake_case = model(**lowerCamelCase , training=lowerCamelCase ) # verify the logits _snake_case = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) _snake_case = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , lowerCamelCase , atol=1e-4 )
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE__ , 2 ) - a def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 2 * x def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2.0 while start <= a: _snake_case = math.pow(SCREAMING_SNAKE_CASE__ , 2 ) return start def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 99_99 , SCREAMING_SNAKE_CASE__ = 0.00000000000001 ): '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _snake_case = get_initial_point(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = value _snake_case = value - fx(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) / fx_derivative(SCREAMING_SNAKE_CASE__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
672
1
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[int] = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = '''git_vision_model''' def __init__( self , lowerCamelCase=768 , lowerCamelCase=3_072 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=16 , lowerCamelCase="quick_gelu" , lowerCamelCase=1e-5 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = num_channels _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = '''git''' def __init__( self , lowerCamelCase=None , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=6 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1_024 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=101 , lowerCamelCase=102 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , pad_token_id=lowerCamelCase , **lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) _snake_case = GitVisionConfig(**lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = num_image_with_embedding _snake_case = bos_token_id _snake_case = eos_token_id def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.__class__.model_type return output
672
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin __magic_name__ : int = False @skip_mps class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = StableDiffusionAttendAndExcitePipeline UpperCAmelCase__ : Any = False UpperCAmelCase__ : Optional[int] = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__ : Dict = TEXT_TO_IMAGE_BATCH_PARAMS.union({'''token_indices'''} ) UpperCAmelCase__ : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase__ : Any = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def UpperCamelCase( cls ): super().setUpClass() torch.use_deterministic_algorithms(lowerCamelCase ) @classmethod def UpperCamelCase( cls ): super().tearDownClass() torch.use_deterministic_algorithms(lowerCamelCase ) def UpperCamelCase( self ): torch.manual_seed(0 ) _snake_case = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase , ) _snake_case = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCamelCase , set_alpha_to_one=lowerCamelCase , ) torch.manual_seed(0 ) _snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="gelu" , projection_dim=512 , ) _snake_case = CLIPTextModel(lowerCamelCase ) _snake_case = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=0 ): if str(lowerCamelCase ).startswith("mps" ): _snake_case = torch.manual_seed(lowerCamelCase ) else: _snake_case = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) _snake_case = _snake_case = { "prompt": "a cat and a frog", "token_indices": [2, 5], "generator": generator, "num_inference_steps": 1, "guidance_scale": 6.0, "output_type": "numpy", "max_iter_to_alter": 2, "thresholds": {0: 0.7}, } return inputs def UpperCamelCase( self ): _snake_case = "cpu" _snake_case = self.get_dummy_components() _snake_case = self.pipeline_class(**lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs(lowerCamelCase ) _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) _snake_case = np.array( [0.63905364, 0.62897307, 0.48599017, 0.5133624, 0.5550048, 0.45769516, 0.50326973, 0.5023139, 0.45384496] ) _snake_case = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase , 1e-3 ) def UpperCamelCase( self ): super().test_cpu_offload_forward_pass(expected_max_diff=5e-4 ) def UpperCamelCase( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def UpperCamelCase( self ): self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7e-4 ) def UpperCamelCase( self ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def UpperCamelCase( self ): super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5e-4 ) def UpperCamelCase( self ): super().test_save_load_local(expected_max_difference=5e-4 ) def UpperCamelCase( self ): super().test_save_load_optional_components(expected_max_difference=4e-4 ) @require_torch_gpu @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @classmethod def UpperCamelCase( cls ): super().setUpClass() torch.use_deterministic_algorithms(lowerCamelCase ) @classmethod def UpperCamelCase( cls ): super().tearDownClass() torch.use_deterministic_algorithms(lowerCamelCase ) def UpperCamelCase( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase( self ): _snake_case = torch.manual_seed(51 ) _snake_case = StableDiffusionAttendAndExcitePipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , safety_checker=lowerCamelCase , torch_dtype=torch.floataa ) pipe.to("cuda" ) _snake_case = "a painting of an elephant with glasses" _snake_case = [5, 7] _snake_case = pipe( prompt=lowerCamelCase , token_indices=lowerCamelCase , guidance_scale=7.5 , generator=lowerCamelCase , num_inference_steps=5 , max_iter_to_alter=5 , output_type="numpy" , ).images[0] _snake_case = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy" ) assert np.abs((expected_image - image).max() ) < 5e-1
672
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __magic_name__ : Dict = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , np.ndarray ): return list(tensor.shape ) _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) if tensor.shape == tf.TensorShape(SCREAMING_SNAKE_CASE__ ): return dynamic _snake_case = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(SCREAMING_SNAKE_CASE__ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise NotImplementedError("Only 1D weight and bias tensors are supported for now, with only a single axis." ) # Get mean and variance on the axis to be normalized _snake_case , _snake_case = tf.nn.moments(SCREAMING_SNAKE_CASE__ , axes=[axis] , keepdims=SCREAMING_SNAKE_CASE__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis _snake_case = [1] * inputs.shape.rank _snake_case = shape_list(SCREAMING_SNAKE_CASE__ )[axis] _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Compute layer normalization using the batch_normalization # function. _snake_case = tf.nn.batch_normalization( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , offset=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , variance_epsilon=SCREAMING_SNAKE_CASE__ , ) return outputs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) _snake_case = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) _snake_case = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ): _snake_case = tf.convert_to_tensor(SCREAMING_SNAKE_CASE__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: _snake_case = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: _snake_case = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) _snake_case = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( SCREAMING_SNAKE_CASE__ , tf.cast(SCREAMING_SNAKE_CASE__ , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(SCREAMING_SNAKE_CASE__ )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. _snake_case = [x for x in data if len(SCREAMING_SNAKE_CASE__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( "The following attributes cannot be saved to HDF5 file because " f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) _snake_case = np.asarray(SCREAMING_SNAKE_CASE__ ) _snake_case = 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(SCREAMING_SNAKE_CASE__ ): _snake_case = chunk_data else: _snake_case = data def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if name in group.attrs: _snake_case = [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs[name]] else: _snake_case = [] _snake_case = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs["%s%d" % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _expand_single_ad_tensor(SCREAMING_SNAKE_CASE__ ): if isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(SCREAMING_SNAKE_CASE__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , SCREAMING_SNAKE_CASE__ )
672
1
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase( self ): _snake_case = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) _snake_case = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house _snake_case = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim _snake_case = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): _snake_case = model(lowerCamelCase )["last_hidden_state"].detach() self.assertEqual(output.shape , lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , lowerCamelCase , atol=1e-3 ) ) @slow def UpperCamelCase( self ): _snake_case = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) _snake_case = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house _snake_case = torch.Size((1, 12, 1_024) ) # batch_size, sequence_length, embedding_vector_dim _snake_case = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): _snake_case = model(lowerCamelCase )["last_hidden_state"].detach() self.assertEqual(output.shape , lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , lowerCamelCase , atol=1e-3 ) )
672
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
1
'''simple docstring''' import re def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return [char.split() for char in re.split(r"[^ a-z A-Z 0-9 \s]" , str_ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' try: _snake_case = split_input(SCREAMING_SNAKE_CASE__ ) if upper: _snake_case = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: _snake_case = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return to_simple_case(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' try: _snake_case = to_simple_case(SCREAMING_SNAKE_CASE__ ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return to_complex_case(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , "_" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return to_complex_case(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , "-" ) if __name__ == "__main__": __import__("""doctest""").testmod()
672
'''simple docstring''' from torch import nn def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'''Unsupported activation function: {act_fn}''' )
672
1
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING __magic_name__ : int = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """salesforce/blip2-opt-2.7b""": """https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = '''blip_2_vision_model''' def __init__( self , lowerCamelCase=1_408 , lowerCamelCase=6_144 , lowerCamelCase=39 , lowerCamelCase=16 , lowerCamelCase=224 , lowerCamelCase=14 , lowerCamelCase="gelu" , lowerCamelCase=0.00001 , lowerCamelCase=0.0 , lowerCamelCase=1e-10 , lowerCamelCase=True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act _snake_case = qkv_bias @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : str = '''blip_2_qformer''' def __init__( self , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=2 , lowerCamelCase=1_408 , **lowerCamelCase , ): super().__init__(pad_token_id=lowerCamelCase , **lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = cross_attention_frequency _snake_case = encoder_hidden_size @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": _snake_case = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = '''blip-2''' UpperCAmelCase__ : str = True def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=32 , **lowerCamelCase ): super().__init__(**lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: _snake_case = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: _snake_case = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) _snake_case = BlipaVisionConfig(**lowerCamelCase ) _snake_case = BlipaQFormerConfig(**lowerCamelCase ) _snake_case = text_config["model_type"] if "model_type" in text_config else "opt" _snake_case = CONFIG_MAPPING[text_model_type](**lowerCamelCase ) _snake_case = self.text_config.tie_word_embeddings _snake_case = self.text_config.is_encoder_decoder _snake_case = num_query_tokens _snake_case = self.vision_config.hidden_size _snake_case = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES _snake_case = 1.0 _snake_case = 0.02 @classmethod def UpperCamelCase( cls , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **lowerCamelCase , ) def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.qformer_config.to_dict() _snake_case = self.text_config.to_dict() _snake_case = self.__class__.model_type return output
672
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __magic_name__ : Tuple = 0 __magic_name__ : Dict = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __magic_name__ : str = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __magic_name__ : Dict = tuple[int, int] class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = pos_x _snake_case = pos_y _snake_case = (pos_y, pos_x) _snake_case = goal_x _snake_case = goal_y _snake_case = g_cost _snake_case = parent _snake_case = self.calculate_heuristic() _snake_case = self.g_cost + self.h_cost def UpperCamelCase( self ): _snake_case = self.pos_x - self.goal_x _snake_case = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCamelCase ) + abs(lowerCamelCase ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowerCamelCase ): return self.f_cost < other.f_cost class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCamelCase ) _snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , lowerCamelCase ) _snake_case = [self.start] _snake_case = [] _snake_case = False def UpperCamelCase( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _snake_case = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCamelCase ) self.closed_nodes.append(lowerCamelCase ) _snake_case = self.get_successors(lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = self.open_nodes.pop(self.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCamelCase ) else: self.open_nodes.append(lowerCamelCase ) return [self.start.pos] def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for action in delta: _snake_case = parent.pos_x + action[1] _snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCamelCase , lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCamelCase , ) ) return successors def UpperCamelCase( self , lowerCamelCase ): _snake_case = node _snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _snake_case = current_node.parent path.reverse() return path class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = False def UpperCamelCase( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _snake_case = self.fwd_astar.open_nodes.pop(0 ) _snake_case = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCamelCase , lowerCamelCase ) self.fwd_astar.closed_nodes.append(lowerCamelCase ) self.bwd_astar.closed_nodes.append(lowerCamelCase ) _snake_case = current_bwd_node _snake_case = current_fwd_node _snake_case = { self.fwd_astar: self.fwd_astar.get_successors(lowerCamelCase ), self.bwd_astar: self.bwd_astar.get_successors(lowerCamelCase ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = astar.open_nodes.pop( astar.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCamelCase ) else: astar.open_nodes.append(lowerCamelCase ) return [self.fwd_astar.start.pos] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self.fwd_astar.retrace_path(lowerCamelCase ) _snake_case = self.bwd_astar.retrace_path(lowerCamelCase ) bwd_path.pop() bwd_path.reverse() _snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __magic_name__ : Optional[int] = (0, 0) __magic_name__ : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __magic_name__ : Any = time.time() __magic_name__ : Optional[int] = AStar(init, goal) __magic_name__ : str = a_star.search() __magic_name__ : List[Any] = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') __magic_name__ : List[str] = time.time() __magic_name__ : Optional[Any] = BidirectionalAStar(init, goal) __magic_name__ : Optional[int] = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
672
1
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ : int = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __magic_name__ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @staticmethod @abstractmethod def UpperCamelCase( lowerCamelCase ): raise NotImplementedError() @abstractmethod def UpperCamelCase( self ): raise NotImplementedError()
672
'''simple docstring''' import string def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "" for i in sequence: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if 65 <= extract <= 90: output += chr(1_55 - extract ) elif 97 <= extract <= 1_22: output += chr(2_19 - extract ) else: output += i return output def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = string.ascii_letters _snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(SCREAMING_SNAKE_CASE__ )] if c in letters else c for c in sequence ) def snake_case_ ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) _snake_case = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'''> atbash_slow(): {timeit("atbash_slow(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) print(f'''> atbash(): {timeit("atbash(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'{example} encrypted in atbash: {atbash(example)}') benchmark()
672
1
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __magic_name__ : Optional[Any] = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , ): _snake_case = [file for file in os.listdir(lowerCamelCase ) if os.path.isfile(os.path.join(lowerCamelCase , lowerCamelCase ) )] if identifier is not None: _snake_case = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowerCamelCase , lowerCamelCase ): for n_ in n_identifier: _snake_case = [file for file in files if n_ not in file] else: _snake_case = [file for file in files if n_identifier not in file] _snake_case = ignore_files or [] ignore_files.append("__init__.py" ) _snake_case = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowerCamelCase ) if only_modules: _snake_case = file.split("." )[0] try: _snake_case = getattr(lowerCamelCase , lowerCamelCase ) _snake_case = doctest.DocTestSuite(lowerCamelCase ) _snake_case = unittest.TextTestRunner().run(lowerCamelCase ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F'''{module_identifier} is not a module.''' ) else: _snake_case = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase( self ): _snake_case = Path("src/transformers" ) _snake_case = "modeling" _snake_case = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowerCamelCase , identifier=lowerCamelCase , ignore_files=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = Path("src/transformers" ) _snake_case = "tokenization" self.analyze_directory(lowerCamelCase , identifier=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = Path("src/transformers" ) _snake_case = "configuration" self.analyze_directory(lowerCamelCase , identifier=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = Path("src/transformers" ) _snake_case = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowerCamelCase , n_identifier=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = Path("docs/source" ) _snake_case = ["favicon.ico"] self.analyze_directory(lowerCamelCase , ignore_files=lowerCamelCase , only_modules=lowerCamelCase )
672
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
1
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Any = logging.get_logger(__name__) __magic_name__ : int = { """facebook/encodec_24khz""": """https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json""", """facebook/encodec_48khz""": """https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = '''encodec''' def __init__( self , lowerCamelCase=[1.5, 3.0, 6.0, 12.0, 24.0] , lowerCamelCase=24_000 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=128 , lowerCamelCase=32 , lowerCamelCase=1 , lowerCamelCase=[8, 5, 4, 2] , lowerCamelCase="weight_norm" , lowerCamelCase=7 , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=2 , lowerCamelCase=True , lowerCamelCase="reflect" , lowerCamelCase=2 , lowerCamelCase=2 , lowerCamelCase=1.0 , lowerCamelCase=1_024 , lowerCamelCase=None , lowerCamelCase=True , **lowerCamelCase , ): _snake_case = target_bandwidths _snake_case = sampling_rate _snake_case = audio_channels _snake_case = normalize _snake_case = chunk_length_s _snake_case = overlap _snake_case = hidden_size _snake_case = num_filters _snake_case = num_residual_layers _snake_case = upsampling_ratios _snake_case = norm_type _snake_case = kernel_size _snake_case = last_kernel_size _snake_case = residual_kernel_size _snake_case = dilation_growth_rate _snake_case = use_causal_conv _snake_case = pad_mode _snake_case = compress _snake_case = num_lstm_layers _snake_case = trim_right_ratio _snake_case = codebook_size _snake_case = codebook_dim if codebook_dim is not None else hidden_size _snake_case = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F'''self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}''' ) super().__init__(**lowerCamelCase ) @property def UpperCamelCase( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def UpperCamelCase( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def UpperCamelCase( self ): _snake_case = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def UpperCamelCase( self ): return int(1_000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
672
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = DistilBertTokenizer UpperCAmelCase__ : Union[str, Any] = DistilBertTokenizerFast UpperCAmelCase__ : List[str] = True @slow def UpperCamelCase( self ): _snake_case = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _snake_case = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
672
1
'''simple docstring''' from collections import deque def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = len(SCREAMING_SNAKE_CASE__ ) _snake_case = deque() _snake_case = [False for _ in range(SCREAMING_SNAKE_CASE__ )] _snake_case = [-1 for _ in range(SCREAMING_SNAKE_CASE__ )] _snake_case = index_of[:] def strong_connect(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = index # the number when this node is seen _snake_case = index # lowest rank node reachable from here index += 1 stack.append(SCREAMING_SNAKE_CASE__ ) _snake_case = True for w in g[v]: if index_of[w] == -1: _snake_case = strong_connect(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: _snake_case = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: _snake_case = [] _snake_case = stack.pop() _snake_case = False component.append(SCREAMING_SNAKE_CASE__ ) while w != v: _snake_case = stack.pop() _snake_case = False component.append(SCREAMING_SNAKE_CASE__ ) components.append(SCREAMING_SNAKE_CASE__ ) return index _snake_case = [] for v in range(SCREAMING_SNAKE_CASE__ ): if index_of[v] == -1: strong_connect(SCREAMING_SNAKE_CASE__ , 0 , SCREAMING_SNAKE_CASE__ ) return components def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [[] for _ in range(SCREAMING_SNAKE_CASE__ )] for u, v in edges: g[u].append(SCREAMING_SNAKE_CASE__ ) return g if __name__ == "__main__": # Test __magic_name__ : List[Any] = 7 __magic_name__ : List[str] = [0, 0, 1, 2, 3, 3, 4, 4, 6] __magic_name__ : Union[str, Any] = [1, 3, 2, 0, 1, 4, 5, 6, 5] __magic_name__ : Dict = [(u, v) for u, v in zip(source, target)] __magic_name__ : List[Any] = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __magic_name__ : Optional[int] = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[int] = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from math import isqrt, loga def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = False return [i for i in range(2 , SCREAMING_SNAKE_CASE__ ) if is_prime[i]] def snake_case_ ( SCREAMING_SNAKE_CASE__ = 80_08_00 , SCREAMING_SNAKE_CASE__ = 80_08_00 ): '''simple docstring''' _snake_case = degree * loga(SCREAMING_SNAKE_CASE__ ) _snake_case = int(SCREAMING_SNAKE_CASE__ ) _snake_case = calculate_prime_numbers(SCREAMING_SNAKE_CASE__ ) _snake_case = 0 _snake_case = 0 _snake_case = len(SCREAMING_SNAKE_CASE__ ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F'{solution() = }')
672
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : Union[str, Any] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = "backbone." if is_semantic else "" _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', "beit.embeddings.cls_token"), (f'''{prefix}patch_embed.proj.weight''', "beit.embeddings.patch_embeddings.projection.weight"), (f'''{prefix}patch_embed.proj.bias''', "beit.embeddings.patch_embeddings.projection.bias"), (f'''{prefix}pos_embed''', "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): _snake_case = "backbone." if is_semantic else "" # queries, keys and values _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = q_bias _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) _snake_case = gamma_a _snake_case = gamma_a def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = False if "rvlcdip" in checkpoint_url else True _snake_case = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE__ , use_mask_token=SCREAMING_SNAKE_CASE__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _snake_case = 10_24 _snake_case = 40_96 _snake_case = 24 _snake_case = 16 # labels if "rvlcdip" in checkpoint_url: _snake_case = 16 _snake_case = "huggingface/label-files" _snake_case = "rvlcdip-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _snake_case = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="cpu" )["model"] _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) # load HuggingFace model _snake_case = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image _snake_case = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ ) _snake_case = prepare_img() _snake_case = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) _snake_case = outputs.logits # verify logits _snake_case = [1, 16] if "rvlcdip" in checkpoint_url else [1, 1_96, 81_92] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE__ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: if has_lm_head: _snake_case = "dit-base" if "base" in checkpoint_url else "dit-large" else: _snake_case = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) if __name__ == "__main__": __magic_name__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) __magic_name__ : Dict = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
672
1
'''simple docstring''' class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self ): _snake_case = 0 _snake_case = 0 _snake_case = {} def UpperCamelCase( self , lowerCamelCase ): if vertex not in self.adjacency: _snake_case = {} self.num_vertices += 1 def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): self.add_vertex(lowerCamelCase ) self.add_vertex(lowerCamelCase ) if head == tail: return _snake_case = weight _snake_case = weight def UpperCamelCase( self ): _snake_case = self.get_edges() for edge in edges: _snake_case , _snake_case , _snake_case = edge edges.remove((tail, head, weight) ) for i in range(len(lowerCamelCase ) ): _snake_case = list(edges[i] ) edges.sort(key=lambda lowerCamelCase : e[2] ) for i in range(len(lowerCamelCase ) - 1 ): if edges[i][2] >= edges[i + 1][2]: _snake_case = edges[i][2] + 1 for edge in edges: _snake_case , _snake_case , _snake_case = edge _snake_case = weight _snake_case = weight def __str__( self ): _snake_case = "" for tail in self.adjacency: for head in self.adjacency[tail]: _snake_case = self.adjacency[head][tail] string += F'''{head} -> {tail} == {weight}\n''' return string.rstrip("\n" ) def UpperCamelCase( self ): _snake_case = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def UpperCamelCase( self ): return self.adjacency.keys() @staticmethod def UpperCamelCase( lowerCamelCase=None , lowerCamelCase=None ): _snake_case = Graph() if vertices is None: _snake_case = [] if edges is None: _snake_case = [] for vertex in vertices: g.add_vertex(lowerCamelCase ) for edge in edges: g.add_edge(*lowerCamelCase ) return g class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self ): _snake_case = {} _snake_case = {} def __len__( self ): return len(self.parent ) def UpperCamelCase( self , lowerCamelCase ): if item in self.parent: return self.find(lowerCamelCase ) _snake_case = item _snake_case = 0 return item def UpperCamelCase( self , lowerCamelCase ): if item not in self.parent: return self.make_set(lowerCamelCase ) if item != self.parent[item]: _snake_case = self.find(self.parent[item] ) return self.parent[item] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self.find(lowerCamelCase ) _snake_case = self.find(lowerCamelCase ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: _snake_case = roota return roota if self.rank[roota] < self.rank[roota]: _snake_case = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 _snake_case = roota return roota return None @staticmethod def UpperCamelCase( lowerCamelCase ): _snake_case = graph.num_vertices _snake_case = Graph.UnionFind() _snake_case = [] while num_components > 1: _snake_case = {} for vertex in graph.get_vertices(): _snake_case = -1 _snake_case = graph.get_edges() for edge in edges: _snake_case , _snake_case , _snake_case = edge edges.remove((tail, head, weight) ) for edge in edges: _snake_case , _snake_case , _snake_case = edge _snake_case = union_find.find(lowerCamelCase ) _snake_case = union_find.find(lowerCamelCase ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: _snake_case = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: _snake_case = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: _snake_case , _snake_case , _snake_case = cheap_edge[vertex] if union_find.find(lowerCamelCase ) != union_find.find(lowerCamelCase ): union_find.union(lowerCamelCase , lowerCamelCase ) mst_edges.append(cheap_edge[vertex] ) _snake_case = num_components - 1 _snake_case = Graph.build(edges=lowerCamelCase ) return mst
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = factor * value _snake_case = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
672
1
'''simple docstring''' import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig __magic_name__ : Any = logging.get_logger(__name__) __magic_name__ : Optional[int] = """T5Config""" def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = jnp.zeros_like(SCREAMING_SNAKE_CASE__ ) _snake_case = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) _snake_case = shifted_input_ids.at[:, 0].set(SCREAMING_SNAKE_CASE__ ) _snake_case = jnp.where(shifted_input_ids == -1_00 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return shifted_input_ids class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = '''mt5''' UpperCAmelCase__ : Optional[int] = MTaConfig class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = '''mt5''' UpperCAmelCase__ : Tuple = MTaConfig class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = '''mt5''' UpperCAmelCase__ : Any = MTaConfig
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __magic_name__ : Optional[int] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"shortest_edge": 256} _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) _snake_case = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging __magic_name__ : List[str] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Dict = ['''input_values''', '''padding_mask'''] def __init__( self , lowerCamelCase = 1 , lowerCamelCase = 24_000 , lowerCamelCase = 0.0 , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(feature_size=lowerCamelCase , sampling_rate=lowerCamelCase , padding_value=lowerCamelCase , **lowerCamelCase ) _snake_case = chunk_length_s _snake_case = overlap @property def UpperCamelCase( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def UpperCamelCase( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if padding and truncation: raise ValueError("Both padding and truncation were set. Make sure you only set one." ) elif padding is None: # by default let's pad the inputs _snake_case = True _snake_case = bool( isinstance(lowerCamelCase , (list, tuple) ) and (isinstance(raw_audio[0] , (np.ndarray, tuple, list) )) ) if is_batched: _snake_case = [np.asarray(lowerCamelCase , dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(lowerCamelCase , np.ndarray ): _snake_case = np.asarray(lowerCamelCase , dtype=np.floataa ) elif isinstance(lowerCamelCase , np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): _snake_case = raw_audio.astype(np.floataa ) # always return batch if not is_batched: _snake_case = [np.asarray(lowerCamelCase ).T] # verify inputs are valid for idx, example in enumerate(lowerCamelCase ): if example.ndim > 2: raise ValueError(F'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(F'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(F'''Expected stereo audio but example has {example.shape[-1]} channels''' ) _snake_case = None _snake_case = BatchFeature({"input_values": raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: _snake_case = min(array.shape[0] for array in raw_audio ) _snake_case = int(np.floor(max_length / self.chunk_stride ) ) _snake_case = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: _snake_case = max(array.shape[0] for array in raw_audio ) _snake_case = int(np.ceil(max_length / self.chunk_stride ) ) _snake_case = (nb_step - 1) * self.chunk_stride + self.chunk_length _snake_case = "max_length" else: _snake_case = input_values # normal padding on batch if padded_inputs is None: _snake_case = self.pad( lowerCamelCase , max_length=lowerCamelCase , truncation=lowerCamelCase , padding=lowerCamelCase , return_attention_mask=lowerCamelCase , ) if padding: _snake_case = padded_inputs.pop("attention_mask" ) _snake_case = [] for example in padded_inputs.pop("input_values" ): if self.feature_size == 1: _snake_case = example[..., None] input_values.append(example.T ) _snake_case = input_values if return_tensors is not None: _snake_case = padded_inputs.convert_to_tensors(lowerCamelCase ) return padded_inputs
672
'''simple docstring''' import baseaa def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaadecode(SCREAMING_SNAKE_CASE__ ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ = 10_00 ): '''simple docstring''' return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
672
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
672
1
'''simple docstring''' import argparse import hashlib # hashlib is only used inside the Test class import struct class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase ): _snake_case = data _snake_case = [0x6745_2301, 0xEFCD_AB89, 0x98BA_DCFE, 0x1032_5476, 0xC3D2_E1F0] @staticmethod def UpperCamelCase( lowerCamelCase , lowerCamelCase ): return ((n << b) | (n >> (32 - b))) & 0xFFFF_FFFF def UpperCamelCase( self ): _snake_case = b"\x80" + b"\x00" * (63 - (len(self.data ) + 8) % 64) _snake_case = self.data + padding + struct.pack(">Q" , 8 * len(self.data ) ) return padded_data def UpperCamelCase( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def UpperCamelCase( self , lowerCamelCase ): _snake_case = list(struct.unpack(">16L" , lowerCamelCase ) ) + [0] * 64 for i in range(16 , 80 ): _snake_case = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def UpperCamelCase( self ): _snake_case = self.padding() _snake_case = self.split_blocks() for block in self.blocks: _snake_case = self.expand_block(lowerCamelCase ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = self.h for i in range(0 , 80 ): if 0 <= i < 20: _snake_case = (b & c) | ((~b) & d) _snake_case = 0x5A82_7999 elif 20 <= i < 40: _snake_case = b ^ c ^ d _snake_case = 0x6ED9_EBA1 elif 40 <= i < 60: _snake_case = (b & c) | (b & d) | (c & d) _snake_case = 0x8F1B_BCDC elif 60 <= i < 80: _snake_case = b ^ c ^ d _snake_case = 0xCA62_C1D6 _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = ( self.rotate(lowerCamelCase , 5 ) + f + e + k + expanded_block[i] & 0xFFFF_FFFF, a, self.rotate(lowerCamelCase , 30 ), c, d, ) _snake_case = ( self.h[0] + a & 0xFFFF_FFFF, self.h[1] + b & 0xFFFF_FFFF, self.h[2] + c & 0xFFFF_FFFF, self.h[3] + d & 0xFFFF_FFFF, self.h[4] + e & 0xFFFF_FFFF, ) return ("{:08x}" * 5).format(*self.h ) def snake_case_ ( ): '''simple docstring''' _snake_case = b"Test String" assert SHAaHash(SCREAMING_SNAKE_CASE__ ).final_hash() == hashlib.shaa(SCREAMING_SNAKE_CASE__ ).hexdigest() # noqa: S324 def snake_case_ ( ): '''simple docstring''' _snake_case = argparse.ArgumentParser(description="Process some strings or files" ) parser.add_argument( "--string" , dest="input_string" , default="Hello World!! Welcome to Cryptography" , help="Hash the string" , ) parser.add_argument("--file" , dest="input_file" , help="Hash contents of a file" ) _snake_case = parser.parse_args() _snake_case = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , "rb" ) as f: _snake_case = f.read() else: _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) print(SHAaHash(SCREAMING_SNAKE_CASE__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
672
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = ['''image_processor'''] UpperCAmelCase__ : Union[str, Any] = '''SamImageProcessor''' def __init__( self , lowerCamelCase ): super().__init__(lowerCamelCase ) _snake_case = self.image_processor _snake_case = -10 _snake_case = self.image_processor.size["longest_edge"] def __call__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = self.image_processor( lowerCamelCase , return_tensors=lowerCamelCase , **lowerCamelCase , ) # pop arguments that are not used in the foward but used nevertheless _snake_case = encoding_image_processor["original_sizes"] if hasattr(lowerCamelCase , "numpy" ): # Checks if Torch or TF tensor _snake_case = original_sizes.numpy() _snake_case , _snake_case , _snake_case = self._check_and_preprocess_points( input_points=lowerCamelCase , input_labels=lowerCamelCase , input_boxes=lowerCamelCase , ) _snake_case = self._normalize_and_convert( lowerCamelCase , lowerCamelCase , input_points=lowerCamelCase , input_labels=lowerCamelCase , input_boxes=lowerCamelCase , return_tensors=lowerCamelCase , ) return encoding_image_processor def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="pt" , ): if input_points is not None: if len(lowerCamelCase ) != len(lowerCamelCase ): _snake_case = [ self._normalize_coordinates(self.target_size , lowerCamelCase , original_sizes[0] ) for point in input_points ] else: _snake_case = [ self._normalize_coordinates(self.target_size , lowerCamelCase , lowerCamelCase ) for point, original_size in zip(lowerCamelCase , lowerCamelCase ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _snake_case , _snake_case = self._pad_points_and_labels(lowerCamelCase , lowerCamelCase ) _snake_case = np.array(lowerCamelCase ) if input_labels is not None: _snake_case = np.array(lowerCamelCase ) if input_boxes is not None: if len(lowerCamelCase ) != len(lowerCamelCase ): _snake_case = [ self._normalize_coordinates(self.target_size , lowerCamelCase , original_sizes[0] , is_bounding_box=lowerCamelCase ) for box in input_boxes ] else: _snake_case = [ self._normalize_coordinates(self.target_size , lowerCamelCase , lowerCamelCase , is_bounding_box=lowerCamelCase ) for box, original_size in zip(lowerCamelCase , lowerCamelCase ) ] _snake_case = np.array(lowerCamelCase ) if input_boxes is not None: if return_tensors == "pt": _snake_case = torch.from_numpy(lowerCamelCase ) # boxes batch size of 1 by default _snake_case = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _snake_case = tf.convert_to_tensor(lowerCamelCase ) # boxes batch size of 1 by default _snake_case = tf.expand_dims(lowerCamelCase , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": _snake_case = torch.from_numpy(lowerCamelCase ) # point batch size of 1 by default _snake_case = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _snake_case = tf.convert_to_tensor(lowerCamelCase ) # point batch size of 1 by default _snake_case = tf.expand_dims(lowerCamelCase , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": _snake_case = torch.from_numpy(lowerCamelCase ) # point batch size of 1 by default _snake_case = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _snake_case = tf.convert_to_tensor(lowerCamelCase ) # point batch size of 1 by default _snake_case = tf.expand_dims(lowerCamelCase , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = max([point.shape[0] for point in input_points] ) _snake_case = [] for i, point in enumerate(lowerCamelCase ): if point.shape[0] != expected_nb_points: _snake_case = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _snake_case = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowerCamelCase ) _snake_case = processed_input_points return input_points, input_labels def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): _snake_case , _snake_case = original_size _snake_case , _snake_case = self.image_processor._get_preprocess_shape(lowerCamelCase , longest_edge=lowerCamelCase ) _snake_case = deepcopy(lowerCamelCase ).astype(lowerCamelCase ) if is_bounding_box: _snake_case = coords.reshape(-1 , 2 , 2 ) _snake_case = coords[..., 0] * (new_w / old_w) _snake_case = coords[..., 1] * (new_h / old_h) if is_bounding_box: _snake_case = coords.reshape(-1 , 4 ) return coords def UpperCamelCase( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if input_points is not None: if hasattr(lowerCamelCase , "numpy" ): # Checks for TF or Torch tensor _snake_case = input_points.numpy().tolist() if not isinstance(lowerCamelCase , lowerCamelCase ) or not isinstance(input_points[0] , lowerCamelCase ): raise ValueError("Input points must be a list of list of floating points." ) _snake_case = [np.array(lowerCamelCase ) for input_point in input_points] else: _snake_case = None if input_labels is not None: if hasattr(lowerCamelCase , "numpy" ): _snake_case = input_labels.numpy().tolist() if not isinstance(lowerCamelCase , lowerCamelCase ) or not isinstance(input_labels[0] , lowerCamelCase ): raise ValueError("Input labels must be a list of list integers." ) _snake_case = [np.array(lowerCamelCase ) for label in input_labels] else: _snake_case = None if input_boxes is not None: if hasattr(lowerCamelCase , "numpy" ): _snake_case = input_boxes.numpy().tolist() if ( not isinstance(lowerCamelCase , lowerCamelCase ) or not isinstance(input_boxes[0] , lowerCamelCase ) or not isinstance(input_boxes[0][0] , lowerCamelCase ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) _snake_case = [np.array(lowerCamelCase ).astype(np.floataa ) for box in input_boxes] else: _snake_case = None return input_points, input_labels, input_boxes @property def UpperCamelCase( self ): _snake_case = self.image_processor.model_input_names return list(dict.fromkeys(lowerCamelCase ) ) def UpperCamelCase( self , *lowerCamelCase , **lowerCamelCase ): return self.image_processor.post_process_masks(*lowerCamelCase , **lowerCamelCase )
672
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
1
'''simple docstring''' import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=50 , lowerCamelCase=0.02 , lowerCamelCase=True , lowerCamelCase=None , ): _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_input_mask _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = use_labels _snake_case = scope def UpperCamelCase( self ): _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = None if self.use_input_mask: _snake_case = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase( self ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , ) def UpperCamelCase( self ): ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = self.prepare_config_and_inputs() _snake_case = True _snake_case = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): _snake_case = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase ) _snake_case = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): _snake_case = True _snake_case = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , ) _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): _snake_case = True _snake_case = True _snake_case = BertGenerationDecoder(config=lowerCamelCase ).to(lowerCamelCase ).eval() # first forward pass _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , use_cache=lowerCamelCase , ) _snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) _snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] # select random slice _snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() _snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() _snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , *lowerCamelCase , ): _snake_case = BertGenerationDecoder(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase( self ): _snake_case , _snake_case , _snake_case , _snake_case = self.prepare_config_and_inputs() _snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () UpperCAmelCase__ : Dict = (BertGenerationDecoder,) if is_torch_available() else () UpperCAmelCase__ : int = ( {'''feature-extraction''': BertGenerationEncoder, '''text-generation''': BertGenerationDecoder} if is_torch_available() else {} ) def UpperCamelCase( self ): _snake_case = BertGenerationEncoderTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase , hidden_size=37 ) def UpperCamelCase( self ): self.config_tester.run_common_tests() def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case , _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = "bert" self.model_tester.create_and_check_model(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*lowerCamelCase ) def UpperCamelCase( self ): # This regression test was failing with PyTorch < 1.3 ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() _snake_case = None self.model_tester.create_and_check_model_as_decoder( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*lowerCamelCase ) @slow def UpperCamelCase( self ): _snake_case = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(lowerCamelCase ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase( self ): _snake_case = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) _snake_case = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): _snake_case = model(lowerCamelCase )[0] _snake_case = torch.Size([1, 8, 1_024] ) self.assertEqual(output.shape , lowerCamelCase ) _snake_case = torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1e-4 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase( self ): _snake_case = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) _snake_case = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): _snake_case = model(lowerCamelCase )[0] _snake_case = torch.Size([1, 8, 50_358] ) self.assertEqual(output.shape , lowerCamelCase ) _snake_case = torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1e-4 ) )
672
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = text_path elif issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [text_path] _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=("train",) ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for split in splits: _snake_case = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader({"train": text_path} , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader({"train": text_path} , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if split: _snake_case = {split: text_path} else: _snake_case = "train" _snake_case = {"train": text_path, "test": text_path} _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
672
1
'''simple docstring''' import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class __SCREAMING_SNAKE_CASE : '''simple docstring''' UpperCAmelCase__ : Optional[Union[str, Path]] = None UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : Optional[Dict] = None UpperCAmelCase__ : Optional[str] = None UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = False UpperCAmelCase__ : bool = True UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : int = 1 UpperCAmelCase__ : Optional[Union[str, bool]] = None UpperCAmelCase__ : bool = False UpperCAmelCase__ : Optional[Dict] = None UpperCAmelCase__ : Optional[str] = None def UpperCamelCase( self ): return self.__class__(**{k: copy.deepcopy(lowerCamelCase ) for k, v in self.__dict__.items()} )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ : Any = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __magic_name__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class __SCREAMING_SNAKE_CASE : '''simple docstring''' def UpperCamelCase( self , lowerCamelCase ): raise NotImplementedError() def UpperCamelCase( self ): raise NotImplementedError() class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase = False , **lowerCamelCase ): _snake_case = tokenizer _snake_case = skip_prompt _snake_case = decode_kwargs # variables used in the streaming process _snake_case = [] _snake_case = 0 _snake_case = True def UpperCamelCase( self , lowerCamelCase ): if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("TextStreamer only supports batch size 1" ) elif len(value.shape ) > 1: _snake_case = value[0] if self.skip_prompt and self.next_tokens_are_prompt: _snake_case = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) _snake_case = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("\n" ): _snake_case = text[self.print_len :] _snake_case = [] _snake_case = 0 # If the last token is a CJK character, we print the characters. elif len(lowerCamelCase ) > 0 and self._is_chinese_char(ord(text[-1] ) ): _snake_case = text[self.print_len :] self.print_len += len(lowerCamelCase ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: _snake_case = text[self.print_len : text.rfind(" " ) + 1] self.print_len += len(lowerCamelCase ) self.on_finalized_text(lowerCamelCase ) def UpperCamelCase( self ): # Flush the cache, if it exists if len(self.token_cache ) > 0: _snake_case = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) _snake_case = text[self.print_len :] _snake_case = [] _snake_case = 0 else: _snake_case = "" _snake_case = True self.on_finalized_text(lowerCamelCase , stream_end=lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = False ): print(lowerCamelCase , flush=lowerCamelCase , end="" if not stream_end else None ) def UpperCamelCase( self , lowerCamelCase ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4E00 and cp <= 0x9FFF) or (cp >= 0x3400 and cp <= 0x4DBF) # or (cp >= 0x2_0000 and cp <= 0x2_A6DF) # or (cp >= 0x2_A700 and cp <= 0x2_B73F) # or (cp >= 0x2_B740 and cp <= 0x2_B81F) # or (cp >= 0x2_B820 and cp <= 0x2_CEAF) # or (cp >= 0xF900 and cp <= 0xFAFF) or (cp >= 0x2_F800 and cp <= 0x2_FA1F) # ): # return True return False class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase = False , lowerCamelCase = None , **lowerCamelCase ): super().__init__(lowerCamelCase , lowerCamelCase , **lowerCamelCase ) _snake_case = Queue() _snake_case = None _snake_case = timeout def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = False ): self.text_queue.put(lowerCamelCase , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self ): return self def UpperCamelCase( self ): _snake_case = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE__ , 2 ) - a def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 2 * x def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2.0 while start <= a: _snake_case = math.pow(SCREAMING_SNAKE_CASE__ , 2 ) return start def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 99_99 , SCREAMING_SNAKE_CASE__ = 0.00000000000001 ): '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _snake_case = get_initial_point(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = value _snake_case = value - fx(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) / fx_derivative(SCREAMING_SNAKE_CASE__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
672
1
'''simple docstring''' import math def snake_case_ ( ): '''simple docstring''' _snake_case = input("Enter message: " ) _snake_case = int(input(f'''Enter key [2-{len(SCREAMING_SNAKE_CASE__ ) - 1}]: ''' ) ) _snake_case = input("Encryption/Decryption [e/d]: " ) if mode.lower().startswith("e" ): _snake_case = encrypt_message(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) elif mode.lower().startswith("d" ): _snake_case = decrypt_message(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f'''Output:\n{text + "|"}''' ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [""] * key for col in range(SCREAMING_SNAKE_CASE__ ): _snake_case = col while pointer < len(SCREAMING_SNAKE_CASE__ ): cipher_text[col] += message[pointer] pointer += key return "".join(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = math.ceil(len(SCREAMING_SNAKE_CASE__ ) / key ) _snake_case = key _snake_case = (num_cols * num_rows) - len(SCREAMING_SNAKE_CASE__ ) _snake_case = [""] * num_cols _snake_case = 0 _snake_case = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _snake_case = 0 row += 1 return "".join(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
672
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[int] = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = '''git_vision_model''' def __init__( self , lowerCamelCase=768 , lowerCamelCase=3_072 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=16 , lowerCamelCase="quick_gelu" , lowerCamelCase=1e-5 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = num_channels _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = '''git''' def __init__( self , lowerCamelCase=None , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=6 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1_024 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=101 , lowerCamelCase=102 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , pad_token_id=lowerCamelCase , **lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) _snake_case = GitVisionConfig(**lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = num_image_with_embedding _snake_case = bos_token_id _snake_case = eos_token_id def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.__class__.model_type return output
672
1
'''simple docstring''' import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging __magic_name__ : int = logging.get_logger(__name__) __magic_name__ : int = {"""vocab_file""": """vocab.txt"""} __magic_name__ : Any = { """vocab_file""": { """facebook/esm2_t6_8M_UR50D""": """https://huggingface.co/facebook/esm2_t6_8M_UR50D/resolve/main/vocab.txt""", """facebook/esm2_t12_35M_UR50D""": """https://huggingface.co/facebook/esm2_t12_35M_UR50D/resolve/main/vocab.txt""", }, } __magic_name__ : Optional[Any] = { """facebook/esm2_t6_8M_UR50D""": 1_024, """facebook/esm2_t12_35M_UR50D""": 1_024, } def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE__ , "r" ) as f: _snake_case = f.read().splitlines() return [l.strip() for l in lines] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = VOCAB_FILES_NAMES UpperCAmelCase__ : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Any = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCamelCase , lowerCamelCase="<unk>" , lowerCamelCase="<cls>" , lowerCamelCase="<pad>" , lowerCamelCase="<mask>" , lowerCamelCase="<eos>" , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = load_vocab_file(lowerCamelCase ) _snake_case = dict(enumerate(self.all_tokens ) ) _snake_case = {tok: ind for ind, tok in enumerate(self.all_tokens )} _snake_case = unk_token _snake_case = cls_token _snake_case = pad_token _snake_case = mask_token _snake_case = eos_token _snake_case = self.all_tokens self._create_trie(self.unique_no_split_tokens ) def UpperCamelCase( self , lowerCamelCase ): return self._id_to_token.get(lowerCamelCase , self.unk_token ) def UpperCamelCase( self , lowerCamelCase ): return self._token_to_id.get(lowerCamelCase , self._token_to_id.get(self.unk_token ) ) def UpperCamelCase( self , lowerCamelCase , **lowerCamelCase ): return text.split() def UpperCamelCase( self , lowerCamelCase=False ): return len(self._id_to_token ) def UpperCamelCase( self ): return {token: i for i, token in enumerate(self.all_tokens )} def UpperCamelCase( self , lowerCamelCase ): return self._token_to_id.get(lowerCamelCase , self._token_to_id.get(self.unk_token ) ) def UpperCamelCase( self , lowerCamelCase ): return self._id_to_token.get(lowerCamelCase , self.unk_token ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = [self.cls_token_id] _snake_case = [self.eos_token_id] # No sep token in ESM vocabulary if token_ids_a is None: if self.eos_token_id is None: return cls + token_ids_a else: return cls + token_ids_a + sep elif self.eos_token_id is None: raise ValueError("Cannot tokenize multiple sequences when EOS token is not set!" ) return cls + token_ids_a + sep + token_ids_a + sep # Multiple inputs always have an EOS token def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if token in self.all_special_ids else 0 for token in token_ids_a] _snake_case = [1] + ([0] * len(lowerCamelCase )) + [1] if token_ids_a is not None: mask += [0] * len(lowerCamelCase ) + [1] return mask def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = os.path.join(lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + "vocab.txt" ) with open(lowerCamelCase , "w" ) as f: f.write("\n".join(self.all_tokens ) ) return (vocab_file,) @property def UpperCamelCase( self ): return self.get_vocab_size(with_added_tokens=lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = False ): return super()._add_tokens(lowerCamelCase , special_tokens=lowerCamelCase )
672
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __magic_name__ : Dict = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , np.ndarray ): return list(tensor.shape ) _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) if tensor.shape == tf.TensorShape(SCREAMING_SNAKE_CASE__ ): return dynamic _snake_case = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(SCREAMING_SNAKE_CASE__ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise NotImplementedError("Only 1D weight and bias tensors are supported for now, with only a single axis." ) # Get mean and variance on the axis to be normalized _snake_case , _snake_case = tf.nn.moments(SCREAMING_SNAKE_CASE__ , axes=[axis] , keepdims=SCREAMING_SNAKE_CASE__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis _snake_case = [1] * inputs.shape.rank _snake_case = shape_list(SCREAMING_SNAKE_CASE__ )[axis] _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Compute layer normalization using the batch_normalization # function. _snake_case = tf.nn.batch_normalization( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , offset=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , variance_epsilon=SCREAMING_SNAKE_CASE__ , ) return outputs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) _snake_case = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) _snake_case = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ): _snake_case = tf.convert_to_tensor(SCREAMING_SNAKE_CASE__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: _snake_case = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: _snake_case = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) _snake_case = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( SCREAMING_SNAKE_CASE__ , tf.cast(SCREAMING_SNAKE_CASE__ , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(SCREAMING_SNAKE_CASE__ )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. _snake_case = [x for x in data if len(SCREAMING_SNAKE_CASE__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( "The following attributes cannot be saved to HDF5 file because " f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) _snake_case = np.asarray(SCREAMING_SNAKE_CASE__ ) _snake_case = 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(SCREAMING_SNAKE_CASE__ ): _snake_case = chunk_data else: _snake_case = data def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if name in group.attrs: _snake_case = [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs[name]] else: _snake_case = [] _snake_case = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs["%s%d" % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _expand_single_ad_tensor(SCREAMING_SNAKE_CASE__ ): if isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(SCREAMING_SNAKE_CASE__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , SCREAMING_SNAKE_CASE__ )
672
1
'''simple docstring''' import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=32 , lowerCamelCase=3 , lowerCamelCase=4 , lowerCamelCase=[10, 20, 30, 40] , lowerCamelCase=[2, 2, 3, 2] , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=["stage2", "stage3", "stage4"] , lowerCamelCase=[2, 3, 4] , lowerCamelCase=None , ): _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = num_labels _snake_case = initializer_range _snake_case = out_features _snake_case = out_indices _snake_case = scope def UpperCamelCase( self ): _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.num_labels ) _snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase( self ): return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = ConvNextVaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = ConvNextVaForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = ConvNextVaBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _snake_case = None _snake_case = ConvNextVaBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCamelCase( self ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {"pixel_values": pixel_values} return config, inputs_dict def UpperCamelCase( self ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) UpperCAmelCase__ : str = ( {'''feature-extraction''': ConvNextVaModel, '''image-classification''': ConvNextVaForImageClassification} if is_torch_available() else {} ) UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : Any = False UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : Union[str, Any] = False def UpperCamelCase( self ): _snake_case = ConvNextVaModelTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def UpperCamelCase( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase( self ): return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def UpperCamelCase( self ): pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def UpperCamelCase( self ): pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): if not self.model_tester.is_training: return for model_class in self.all_model_classes: _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_with_labels() _snake_case = True if model_class.__name__ in [ *get_values(lowerCamelCase ), *get_values(lowerCamelCase ), ]: continue _snake_case = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.train() _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) _snake_case = model(**lowerCamelCase ).loss loss.backward() def UpperCamelCase( self ): if not self.model_tester.is_training: return for model_class in self.all_model_classes: _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_with_labels() _snake_case = False _snake_case = True if ( model_class.__name__ in [*get_values(lowerCamelCase ), *get_values(lowerCamelCase )] or not model_class.supports_gradient_checkpointing ): continue _snake_case = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.gradient_checkpointing_enable() model.train() _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) _snake_case = model(**lowerCamelCase ).loss loss.backward() def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(lowerCamelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): def check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def UpperCamelCase( self ): for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = ConvNextVaModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' _snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase( self ): return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def UpperCamelCase( self ): _snake_case = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(lowerCamelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = preprocessor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): _snake_case = model(**lowerCamelCase ) # verify the logits _snake_case = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) _snake_case = torch.tensor([0.9996, 0.1966, -0.4386] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1e-4 ) )
672
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
1
'''simple docstring''' from collections import deque class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = process_name # process name _snake_case = arrival_time # arrival time of the process # completion time of finished process or last interrupted time _snake_case = arrival_time _snake_case = burst_time # remaining burst time _snake_case = 0 # total time of the process wait in ready queue _snake_case = 0 # time from arrival time to completion time class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): # total number of mlfq's queues _snake_case = number_of_queues # time slice of queues that round robin algorithm applied _snake_case = time_slices # unfinished process is in this ready_queue _snake_case = queue # current time _snake_case = current_time # finished process is in this sequence queue _snake_case = deque() def UpperCamelCase( self ): _snake_case = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for i in range(len(lowerCamelCase ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for i in range(len(lowerCamelCase ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for i in range(len(lowerCamelCase ) ): completion_times.append(queue[i].stop_time ) return completion_times def UpperCamelCase( self , lowerCamelCase ): return [q.burst_time for q in queue] def UpperCamelCase( self , lowerCamelCase ): process.waiting_time += self.current_time - process.stop_time return process.waiting_time def UpperCamelCase( self , lowerCamelCase ): _snake_case = deque() # sequence deque of finished process while len(lowerCamelCase ) != 0: _snake_case = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(lowerCamelCase ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 _snake_case = 0 # set the process's turnaround time because it is finished _snake_case = self.current_time - cp.arrival_time # set the completion time _snake_case = self.current_time # add the process to queue that has finished queue finished.append(lowerCamelCase ) self.finish_queue.extend(lowerCamelCase ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(lowerCamelCase ) ): _snake_case = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(lowerCamelCase ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time _snake_case = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(lowerCamelCase ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished _snake_case = 0 # set the finish time _snake_case = self.current_time # update the process' turnaround time because it is finished _snake_case = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(lowerCamelCase ) self.finish_queue.extend(lowerCamelCase ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def UpperCamelCase( self ): # all queues except last one have round_robin algorithm for i in range(self.number_of_queues - 1 ): _snake_case , _snake_case = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest __magic_name__ : Tuple = Process("""P1""", 0, 53) __magic_name__ : str = Process("""P2""", 0, 17) __magic_name__ : Dict = Process("""P3""", 0, 68) __magic_name__ : Any = Process("""P4""", 0, 24) __magic_name__ : Any = 3 __magic_name__ : Any = [17, 25] __magic_name__ : Union[str, Any] = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={"""queue""": deque([Pa, Pa, Pa, Pa])}) __magic_name__ : int = Process("""P1""", 0, 53) __magic_name__ : List[Any] = Process("""P2""", 0, 17) __magic_name__ : List[Any] = Process("""P3""", 0, 68) __magic_name__ : Optional[Any] = Process("""P4""", 0, 24) __magic_name__ : Any = 3 __magic_name__ : Union[str, Any] = [17, 25] __magic_name__ : Dict = deque([Pa, Pa, Pa, Pa]) __magic_name__ : Optional[Any] = MLFQ(number_of_queues, time_slices, queue, 0) __magic_name__ : str = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F'waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}' ) # print completion times of processes(P1, P2, P3, P4) print( F'completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}' ) # print total turnaround times of processes(P1, P2, P3, P4) print( F'turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}' ) # print sequence of finished processes print( F'sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}' )
672
'''simple docstring''' from torch import nn def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'''Unsupported activation function: {act_fn}''' )
672
1
'''simple docstring''' __magic_name__ : Union[str, Any] = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] __magic_name__ : Any = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] __magic_name__ : List[str] = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert len(str(SCREAMING_SNAKE_CASE__ ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _snake_case = year // 1_00 _snake_case = (5 * (century % 4) + 2) % 7 _snake_case = year % 1_00 _snake_case = centurian % 12 _snake_case = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _snake_case = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) _snake_case = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __magic_name__ : Tuple = 0 __magic_name__ : Dict = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __magic_name__ : str = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __magic_name__ : Dict = tuple[int, int] class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = pos_x _snake_case = pos_y _snake_case = (pos_y, pos_x) _snake_case = goal_x _snake_case = goal_y _snake_case = g_cost _snake_case = parent _snake_case = self.calculate_heuristic() _snake_case = self.g_cost + self.h_cost def UpperCamelCase( self ): _snake_case = self.pos_x - self.goal_x _snake_case = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCamelCase ) + abs(lowerCamelCase ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowerCamelCase ): return self.f_cost < other.f_cost class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCamelCase ) _snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , lowerCamelCase ) _snake_case = [self.start] _snake_case = [] _snake_case = False def UpperCamelCase( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _snake_case = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCamelCase ) self.closed_nodes.append(lowerCamelCase ) _snake_case = self.get_successors(lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = self.open_nodes.pop(self.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCamelCase ) else: self.open_nodes.append(lowerCamelCase ) return [self.start.pos] def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for action in delta: _snake_case = parent.pos_x + action[1] _snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCamelCase , lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCamelCase , ) ) return successors def UpperCamelCase( self , lowerCamelCase ): _snake_case = node _snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _snake_case = current_node.parent path.reverse() return path class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = False def UpperCamelCase( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _snake_case = self.fwd_astar.open_nodes.pop(0 ) _snake_case = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCamelCase , lowerCamelCase ) self.fwd_astar.closed_nodes.append(lowerCamelCase ) self.bwd_astar.closed_nodes.append(lowerCamelCase ) _snake_case = current_bwd_node _snake_case = current_fwd_node _snake_case = { self.fwd_astar: self.fwd_astar.get_successors(lowerCamelCase ), self.bwd_astar: self.bwd_astar.get_successors(lowerCamelCase ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = astar.open_nodes.pop( astar.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCamelCase ) else: astar.open_nodes.append(lowerCamelCase ) return [self.fwd_astar.start.pos] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self.fwd_astar.retrace_path(lowerCamelCase ) _snake_case = self.bwd_astar.retrace_path(lowerCamelCase ) bwd_path.pop() bwd_path.reverse() _snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __magic_name__ : Optional[int] = (0, 0) __magic_name__ : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __magic_name__ : Any = time.time() __magic_name__ : Optional[int] = AStar(init, goal) __magic_name__ : str = a_star.search() __magic_name__ : List[Any] = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') __magic_name__ : List[str] = time.time() __magic_name__ : Optional[Any] = BidirectionalAStar(init, goal) __magic_name__ : Optional[int] = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
672
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = StableDiffusionSAGPipeline UpperCAmelCase__ : Optional[int] = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__ : List[str] = TEXT_TO_IMAGE_BATCH_PARAMS UpperCAmelCase__ : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase__ : Any = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase__ : Optional[int] = False def UpperCamelCase( self ): torch.manual_seed(0 ) _snake_case = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) _snake_case = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCamelCase , set_alpha_to_one=lowerCamelCase , ) torch.manual_seed(0 ) _snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) _snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) _snake_case = CLIPTextModel(lowerCamelCase ) _snake_case = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=0 ): if str(lowerCamelCase ).startswith("mps" ): _snake_case = torch.manual_seed(lowerCamelCase ) else: _snake_case = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) _snake_case = { "prompt": ".", "generator": generator, "num_inference_steps": 2, "guidance_scale": 1.0, "sag_scale": 1.0, "output_type": "numpy", } return inputs def UpperCamelCase( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase( self ): _snake_case = StableDiffusionSAGPipeline.from_pretrained("CompVis/stable-diffusion-v1-4" ) _snake_case = sag_pipe.to(lowerCamelCase ) sag_pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = "." _snake_case = torch.manual_seed(0 ) _snake_case = sag_pipe( [prompt] , generator=lowerCamelCase , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type="np" ) _snake_case = output.images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.1568, 0.1738, 0.1695, 0.1693, 0.1507, 0.1705, 0.1547, 0.1751, 0.1949] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-2 def UpperCamelCase( self ): _snake_case = StableDiffusionSAGPipeline.from_pretrained("stabilityai/stable-diffusion-2-1-base" ) _snake_case = sag_pipe.to(lowerCamelCase ) sag_pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = "." _snake_case = torch.manual_seed(0 ) _snake_case = sag_pipe( [prompt] , generator=lowerCamelCase , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type="np" ) _snake_case = output.images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.3459, 0.2876, 0.2537, 0.3002, 0.2671, 0.2160, 0.3026, 0.2262, 0.2371] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-2 def UpperCamelCase( self ): _snake_case = StableDiffusionSAGPipeline.from_pretrained("stabilityai/stable-diffusion-2-1-base" ) _snake_case = sag_pipe.to(lowerCamelCase ) sag_pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = "." _snake_case = torch.manual_seed(0 ) _snake_case = sag_pipe( [prompt] , width=768 , height=512 , generator=lowerCamelCase , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type="np" , ) _snake_case = output.images assert image.shape == (1, 512, 768, 3)
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ : int = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __magic_name__ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint __magic_name__ : Union[str, Any] = { """169M""": 12, """430M""": 24, """1B5""": 24, """3B""": 32, """7B""": 32, """14B""": 40, } __magic_name__ : List[str] = { """169M""": 768, """430M""": 1_024, """1B5""": 2_048, """3B""": 2_560, """7B""": 4_096, """14B""": 5_120, } def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = list(state_dict.keys() ) for name in state_dict_keys: _snake_case = state_dict.pop(SCREAMING_SNAKE_CASE__ ) # emb -> embedding if name.startswith("emb." ): _snake_case = name.replace("emb." , "embeddings." ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("blocks.0.ln0" ): _snake_case = name.replace("blocks.0.ln0" , "blocks.0.pre_ln" ) # att -> attention _snake_case = re.sub(r"blocks\.(\d+)\.att" , r"blocks.\1.attention" , SCREAMING_SNAKE_CASE__ ) # ffn -> feed_forward _snake_case = re.sub(r"blocks\.(\d+)\.ffn" , r"blocks.\1.feed_forward" , SCREAMING_SNAKE_CASE__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(".time_mix_k" ): _snake_case = name.replace(".time_mix_k" , ".time_mix_key" ) # time_mix_v -> time_mix_value and reshape if name.endswith(".time_mix_v" ): _snake_case = name.replace(".time_mix_v" , ".time_mix_value" ) # time_mix_r -> time_mix_key and reshape if name.endswith(".time_mix_r" ): _snake_case = name.replace(".time_mix_r" , ".time_mix_receptance" ) if name != "head.weight": _snake_case = "rwkv." + name _snake_case = weight return state_dict def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=None ): '''simple docstring''' if tokenizer_file is None: print("No `--tokenizer_file` provided, we will use the default tokenizer." ) _snake_case = 5_02_77 _snake_case = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-20b" ) else: _snake_case = PreTrainedTokenizerFast(tokenizer_file=SCREAMING_SNAKE_CASE__ ) _snake_case = len(SCREAMING_SNAKE_CASE__ ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) # 2. Build the config _snake_case = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _snake_case = candidate break if size is None: raise ValueError("Could not infer the size, please provide it with the `--size` argument." ) if size not in possible_sizes: raise ValueError(f'''`size` should be one of {possible_sizes}, got {size}.''' ) _snake_case = RwkvConfig( vocab_size=SCREAMING_SNAKE_CASE__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(SCREAMING_SNAKE_CASE__ ) # 3. Download model file then convert state_dict _snake_case = hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = torch.load(SCREAMING_SNAKE_CASE__ , map_location="cpu" ) _snake_case = convert_state_dict(SCREAMING_SNAKE_CASE__ ) # 4. Split in shards and save _snake_case , _snake_case = shard_checkpoint(SCREAMING_SNAKE_CASE__ ) for shard_file, shard in shards.items(): torch.save(SCREAMING_SNAKE_CASE__ , os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) if index is not None: _snake_case = os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Save the index as well with open(SCREAMING_SNAKE_CASE__ , "w" , encoding="utf-8" ) as f: _snake_case = json.dumps(SCREAMING_SNAKE_CASE__ , indent=2 , sort_keys=SCREAMING_SNAKE_CASE__ ) + "\n" f.write(SCREAMING_SNAKE_CASE__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( "Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model." ) _snake_case = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _snake_case = torch.load(os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("Please provide a `model_name` to push the model to the Hub." ) _snake_case = AutoModelForCausalLM.from_pretrained(SCREAMING_SNAKE_CASE__ ) model.push_to_hub(SCREAMING_SNAKE_CASE__ , max_shard_size="2GB" ) tokenizer.push_to_hub(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--repo_id""", default=None, type=str, required=True, help="""Repo ID from which to pull the checkpoint.""" ) parser.add_argument( """--checkpoint_file""", default=None, type=str, required=True, help="""Name of the checkpoint file in the repo.""" ) parser.add_argument( """--output_dir""", default=None, type=str, required=True, help="""Where to save the converted model.""" ) parser.add_argument( """--tokenizer_file""", default=None, type=str, help="""Path to the tokenizer file to use (if not provided, only the model is converted).""", ) parser.add_argument( """--size""", default=None, type=str, help="""Size of the model. Will be inferred from the `checkpoint_file` if not passed.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Push to the Hub the converted model.""", ) parser.add_argument( """--model_name""", default=None, type=str, help="""Name of the pushed model on the Hub, including the username / organization.""", ) __magic_name__ : Any = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
672
'''simple docstring''' import string def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "" for i in sequence: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if 65 <= extract <= 90: output += chr(1_55 - extract ) elif 97 <= extract <= 1_22: output += chr(2_19 - extract ) else: output += i return output def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = string.ascii_letters _snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(SCREAMING_SNAKE_CASE__ )] if c in letters else c for c in sequence ) def snake_case_ ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) _snake_case = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'''> atbash_slow(): {timeit("atbash_slow(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) print(f'''> atbash(): {timeit("atbash(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'{example} encrypted in atbash: {atbash(example)}') benchmark()
672
1
'''simple docstring''' import pprint import requests __magic_name__ : Tuple = """https://zenquotes.io/api""" def snake_case_ ( ): '''simple docstring''' return requests.get(API_ENDPOINT_URL + "/today" ).json() def snake_case_ ( ): '''simple docstring''' return requests.get(API_ENDPOINT_URL + "/random" ).json() if __name__ == "__main__": __magic_name__ : Dict = random_quotes() pprint.pprint(response)
672
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging __magic_name__ : int = logging.get_logger(__name__) __magic_name__ : Tuple = {"""vocab_file""": """spiece.model"""} __magic_name__ : Any = { """vocab_file""": { """t5-small""": """https://huggingface.co/t5-small/resolve/main/spiece.model""", """t5-base""": """https://huggingface.co/t5-base/resolve/main/spiece.model""", """t5-large""": """https://huggingface.co/t5-large/resolve/main/spiece.model""", """t5-3b""": """https://huggingface.co/t5-3b/resolve/main/spiece.model""", """t5-11b""": """https://huggingface.co/t5-11b/resolve/main/spiece.model""", } } # TODO(PVP) - this should be removed in Transformers v5 __magic_name__ : Optional[Any] = { """t5-small""": 512, """t5-base""": 512, """t5-large""": 512, """t5-3b""": 512, """t5-11b""": 512, } __magic_name__ : List[Any] = """▁""" class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = VOCAB_FILES_NAMES UpperCAmelCase__ : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : str = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCamelCase , lowerCamelCase="</s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase=100 , lowerCamelCase=None , lowerCamelCase = None , lowerCamelCase=True , **lowerCamelCase , ): # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: _snake_case = [F'''<extra_id_{i}>''' for i in range(lowerCamelCase )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _snake_case = len(set(filter(lambda lowerCamelCase : bool("extra_id" in str(lowerCamelCase ) ) , lowerCamelCase ) ) ) if extra_tokens != extra_ids: raise ValueError( F'''Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are''' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens" ) if legacy: logger.warning_once( F'''You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to''' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565" ) _snake_case = legacy _snake_case = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowerCamelCase , unk_token=lowerCamelCase , pad_token=lowerCamelCase , extra_ids=lowerCamelCase , additional_special_tokens=lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , legacy=lowerCamelCase , **lowerCamelCase , ) _snake_case = vocab_file _snake_case = extra_ids _snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase ) @staticmethod def UpperCamelCase( lowerCamelCase , lowerCamelCase , lowerCamelCase ): if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: _snake_case = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" F''' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this''' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" F''' {pretrained_model_name_or_path} automatically truncating your input to''' F''' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences''' F''' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with''' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value." , lowerCamelCase , ) return max_model_length @property def UpperCamelCase( self ): return self.sp_model.get_piece_size() + self._extra_ids def UpperCamelCase( self ): _snake_case = {self.convert_ids_to_tokens(lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase , token_ids_a=lowerCamelCase , already_has_special_tokens=lowerCamelCase ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowerCamelCase )) + [1] return ([0] * len(lowerCamelCase )) + [1] + ([0] * len(lowerCamelCase )) + [1] def UpperCamelCase( self ): return list( set(filter(lambda lowerCamelCase : bool(re.search(R"<extra_id_\d+>" , lowerCamelCase ) ) is not None , self.additional_special_tokens ) ) ) def UpperCamelCase( self ): return [self._convert_token_to_id(lowerCamelCase ) for token in self.get_sentinel_tokens()] def UpperCamelCase( self , lowerCamelCase ): if len(lowerCamelCase ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'''This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated''' " eos tokens being added." ) return token_ids else: return token_ids + [self.eos_token_id] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = self._add_eos_if_not_present(lowerCamelCase ) if token_ids_a is None: return token_ids_a else: _snake_case = self._add_eos_if_not_present(lowerCamelCase ) return token_ids_a + token_ids_a def __getstate__( self ): _snake_case = self.__dict__.copy() _snake_case = None return state def __setstate__( self , lowerCamelCase ): _snake_case = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _snake_case = {} _snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase( self , lowerCamelCase , **lowerCamelCase ): # Replace the SPIECE_UNDERLINE with a space to make sure SPIECE_UNDERLINE is only used at # the beginning of the text if not self.legacy: _snake_case = SPIECE_UNDERLINE + text.replace(lowerCamelCase , " " ) return super().tokenize(lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , **lowerCamelCase ): if not self.legacy: _snake_case = text.startswith(lowerCamelCase ) if is_first: _snake_case = text[1:] _snake_case = self.sp_model.encode(lowerCamelCase , out_type=lowerCamelCase ) if not self.legacy and not is_first and not text.startswith(" " ) and tokens[0].startswith(lowerCamelCase ): _snake_case = ([tokens[0][1:]] if len(tokens[0] ) > 1 else []) + tokens[1:] return tokens def UpperCamelCase( self , lowerCamelCase ): if token.startswith("<extra_id_" ): _snake_case = re.match(R"<extra_id_(\d+)>" , lowerCamelCase ) _snake_case = int(match.group(1 ) ) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase ): if index < self.sp_model.get_piece_size(): _snake_case = self.sp_model.IdToPiece(lowerCamelCase ) else: _snake_case = F'''<extra_id_{self.vocab_size - 1 - index}>''' return token def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] _snake_case = "" _snake_case = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCamelCase ) + token _snake_case = True _snake_case = [] else: current_sub_tokens.append(lowerCamelCase ) _snake_case = False out_string += self.sp_model.decode(lowerCamelCase ) return out_string.strip() def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): if not os.path.isdir(lowerCamelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _snake_case = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase , "wb" ) as fi: _snake_case = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase ) return (out_vocab_file,)
672
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
1
'''simple docstring''' import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="session" ) def snake_case_ ( ): '''simple docstring''' _snake_case = 10 _snake_case = datasets.Features( { "tokens": datasets.Sequence(datasets.Value("string" ) ), "labels": datasets.Sequence(datasets.ClassLabel(names=["negative", "positive"] ) ), "answers": datasets.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), "id": datasets.Value("int64" ), } ) _snake_case = datasets.Dataset.from_dict( { "tokens": [["foo"] * 5] * n, "labels": [[1] * 5] * n, "answers": [{"answer_start": [97], "text": ["1976"]}] * 10, "id": list(range(SCREAMING_SNAKE_CASE__ ) ), } , features=SCREAMING_SNAKE_CASE__ , ) return dataset @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "file.arrow" ) dataset.map(cache_file_name=SCREAMING_SNAKE_CASE__ ) return filename # FILE_CONTENT + files __magic_name__ : Optional[int] = """\ Text data. Second line of data.""" @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt" _snake_case = FILE_CONTENT with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return filename @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import bza _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.bz2" _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) with bza.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import gzip _snake_case = str(tmp_path_factory.mktemp("data" ) / "file.txt.gz" ) _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) with gzip.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if datasets.config.LZ4_AVAILABLE: import lza.frame _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.lz4" _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) with lza.frame.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if datasets.config.PY7ZR_AVAILABLE: import pyazr _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.7z" with pyazr.SevenZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as archive: archive.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import tarfile _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import lzma _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.xz" _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) with lzma.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import zipfile _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _snake_case = tmp_path_factory.mktemp("data" ) / "file.txt.zst" _snake_case = bytes(SCREAMING_SNAKE_CASE__ , "utf-8" ) with zstd.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "file.xml" _snake_case = textwrap.dedent( "\\n <?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n <tmx version=\"1.4\">\n <header segtype=\"sentence\" srclang=\"ca\" />\n <body>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return filename __magic_name__ : Dict = [ {"""col_1""": """0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """3""", """col_2""": 3, """col_3""": 3.0}, ] __magic_name__ : Union[str, Any] = [ {"""col_1""": """4""", """col_2""": 4, """col_3""": 4.0}, {"""col_1""": """5""", """col_2""": 5, """col_3""": 5.0}, ] __magic_name__ : List[Any] = { """col_1""": ["""0""", """1""", """2""", """3"""], """col_2""": [0, 1, 2, 3], """col_3""": [0.0, 1.0, 2.0, 3.0], } __magic_name__ : Union[str, Any] = [ {"""col_3""": 0.0, """col_1""": """0""", """col_2""": 0}, {"""col_3""": 1.0, """col_1""": """1""", """col_2""": 1}, ] __magic_name__ : List[str] = [ {"""col_1""": """s0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """s1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """s2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """s3""", """col_2""": 3, """col_3""": 3.0}, ] @pytest.fixture(scope="session" ) def snake_case_ ( ): '''simple docstring''' return DATA_DICT_OF_LISTS @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = datasets.Dataset.from_dict(SCREAMING_SNAKE_CASE__ ) _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.arrow" ) dataset.map(cache_file_name=SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.sqlite" ) with contextlib.closing(sqlitea.connect(SCREAMING_SNAKE_CASE__ ) ) as con: _snake_case = con.cursor() cur.execute("CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)" ) for item in DATA: cur.execute("INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)" , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.csv" ) with open(SCREAMING_SNAKE_CASE__ , "w" , newline="" ) as f: _snake_case = csv.DictWriter(SCREAMING_SNAKE_CASE__ , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset2.csv" ) with open(SCREAMING_SNAKE_CASE__ , "w" , newline="" ) as f: _snake_case = csv.DictWriter(SCREAMING_SNAKE_CASE__ , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import bza _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.csv.bz2" with open(SCREAMING_SNAKE_CASE__ , "rb" ) as f: _snake_case = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(csv_path.replace(".csv" , ".CSV" ) ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(csva_path.replace(".csv" , ".CSV" ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.parquet" ) _snake_case = pa.schema( { "col_1": pa.string(), "col_2": pa.intaa(), "col_3": pa.floataa(), } ) with open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: _snake_case = pq.ParquetWriter(SCREAMING_SNAKE_CASE__ , schema=SCREAMING_SNAKE_CASE__ ) _snake_case = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(SCREAMING_SNAKE_CASE__ ) )] for k in DATA[0]} , schema=SCREAMING_SNAKE_CASE__ ) writer.write_table(SCREAMING_SNAKE_CASE__ ) writer.close() return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) _snake_case = {"data": DATA} with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) _snake_case = {"data": DATA_DICT_OF_LISTS} with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in DATA: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset2.jsonl" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in DATA: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset_312.jsonl" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in DATA_312: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset-str.jsonl" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in DATA_STR: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import gzip _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.txt.gz" ) with open(SCREAMING_SNAKE_CASE__ , "rb" ) as orig_file: with gzip.open(SCREAMING_SNAKE_CASE__ , "wb" ) as zipped_file: zipped_file.writelines(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' import gzip _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl.gz" ) with open(SCREAMING_SNAKE_CASE__ , "rb" ) as orig_file: with gzip.open(SCREAMING_SNAKE_CASE__ , "wb" ) as zipped_file: zipped_file.writelines(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("nested" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) f.add(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("nested" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = ["0", "1", "2", "3"] _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset.txt" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = ["0", "1", "2", "3"] _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset2.txt" ) with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = ["0", "1", "2", "3"] _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.abc" with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.text.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.text.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE__ ) ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.ext.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename("unsupported.ext" ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename("unsupported_2.ext" ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "\n".join(["First", "Second\u2029with Unicode new line", "Third"] ) _snake_case = str(tmp_path_factory.mktemp("data" ) / "dataset_with_unicode_new_lines.txt" ) with open(SCREAMING_SNAKE_CASE__ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE__ ) return path @pytest.fixture(scope="session" ) def snake_case_ ( ): '''simple docstring''' return os.path.join("tests" , "features" , "data" , "test_image_rgb.jpg" ) @pytest.fixture(scope="session" ) def snake_case_ ( ): '''simple docstring''' return os.path.join("tests" , "features" , "data" , "test_audio_44100.wav" ) @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data" ) / "dataset.img.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE__ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ) ) f.write(SCREAMING_SNAKE_CASE__ , arcname=os.path.basename(SCREAMING_SNAKE_CASE__ ).replace(".jpg" , "2.jpg" ) ) return path @pytest.fixture(scope="session" ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path_factory.mktemp("data_dir" ) (data_dir / "subdir").mkdir() with open(data_dir / "subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / "subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden file with open(data_dir / "subdir" / ".test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / ".subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / ".subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) return data_dir
672
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = DistilBertTokenizer UpperCAmelCase__ : Union[str, Any] = DistilBertTokenizerFast UpperCAmelCase__ : List[str] = True @slow def UpperCamelCase( self ): _snake_case = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _snake_case = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
672
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor __magic_name__ : List[str] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def __init__( self , *lowerCamelCase , **lowerCamelCase ): warnings.warn( "The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DonutImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __magic_name__ : Optional[int] = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[int] = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = DistilBertTokenizer UpperCAmelCase__ : Union[str, Any] = DistilBertTokenizerFast UpperCAmelCase__ : List[str] = True @slow def UpperCamelCase( self ): _snake_case = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _snake_case = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
672
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : Union[str, Any] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = "backbone." if is_semantic else "" _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', "beit.embeddings.cls_token"), (f'''{prefix}patch_embed.proj.weight''', "beit.embeddings.patch_embeddings.projection.weight"), (f'''{prefix}patch_embed.proj.bias''', "beit.embeddings.patch_embeddings.projection.bias"), (f'''{prefix}pos_embed''', "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): _snake_case = "backbone." if is_semantic else "" # queries, keys and values _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = q_bias _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) _snake_case = gamma_a _snake_case = gamma_a def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = False if "rvlcdip" in checkpoint_url else True _snake_case = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE__ , use_mask_token=SCREAMING_SNAKE_CASE__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _snake_case = 10_24 _snake_case = 40_96 _snake_case = 24 _snake_case = 16 # labels if "rvlcdip" in checkpoint_url: _snake_case = 16 _snake_case = "huggingface/label-files" _snake_case = "rvlcdip-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _snake_case = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="cpu" )["model"] _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) # load HuggingFace model _snake_case = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image _snake_case = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ ) _snake_case = prepare_img() _snake_case = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) _snake_case = outputs.logits # verify logits _snake_case = [1, 16] if "rvlcdip" in checkpoint_url else [1, 1_96, 81_92] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE__ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: if has_lm_head: _snake_case = "dit-base" if "base" in checkpoint_url else "dit-large" else: _snake_case = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) if __name__ == "__main__": __magic_name__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) __magic_name__ : Dict = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
672
1
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. _snake_case = [[1, 2, 4], [1, 2, 3, 4]] _snake_case = DisjunctiveConstraint(lowerCamelCase ) self.assertTrue(isinstance(dc.token_ids , lowerCamelCase ) ) with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def UpperCamelCase( self ): # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). _snake_case = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint(lowerCamelCase ) # fails here def UpperCamelCase( self ): _snake_case = [[1, 2, 3], [1, 2, 4]] _snake_case = DisjunctiveConstraint(lowerCamelCase ) _snake_case , _snake_case , _snake_case = dc.update(1 ) _snake_case = stepped is True and completed is False and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) _snake_case = stepped is True and completed is False and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(3 ) _snake_case = stepped is True and completed is True and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def UpperCamelCase( self ): _snake_case = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] _snake_case = DisjunctiveConstraint(lowerCamelCase ) _snake_case , _snake_case , _snake_case = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) _snake_case , _snake_case , _snake_case = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() _snake_case , _snake_case , _snake_case = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = factor * value _snake_case = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
672
1
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class __SCREAMING_SNAKE_CASE ( pl.LightningModule ): '''simple docstring''' def __init__( self , lowerCamelCase ): super().__init__() _snake_case = model _snake_case = 2 _snake_case = nn.Linear(self.model.config.hidden_size , self.num_labels ) def UpperCamelCase( self ): pass def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = LongformerModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) _snake_case = LightningModel(SCREAMING_SNAKE_CASE__ ) _snake_case = torch.load(SCREAMING_SNAKE_CASE__ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _snake_case = LongformerForQuestionAnswering.from_pretrained(SCREAMING_SNAKE_CASE__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Conversion successful. Model saved under {pytorch_dump_folder_path}''' ) if __name__ == "__main__": __magic_name__ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __magic_name__ : int = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=[] ): '''simple docstring''' _snake_case = size[0] - overlap_pixels * 2 _snake_case = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels _snake_case = np.ones((size_y, size_x) , dtype=np.uinta ) * 2_55 _snake_case = np.pad(SCREAMING_SNAKE_CASE__ , mode="linear_ramp" , pad_width=SCREAMING_SNAKE_CASE__ , end_values=0 ) if "l" in remove_borders: _snake_case = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: _snake_case = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: _snake_case = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: _snake_case = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return max(SCREAMING_SNAKE_CASE__ , min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = list(SCREAMING_SNAKE_CASE__ ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap _snake_case = clamp_rect(SCREAMING_SNAKE_CASE__ , [0, 0] , [image_size[0], image_size[1]] ) return rect def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = Image.new("RGB" , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(SCREAMING_SNAKE_CASE__ , (original_slice, 0) ) return result def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) _snake_case = tile.crop(SCREAMING_SNAKE_CASE__ ) return tile def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = n % d return n - divisor class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = 350 , ): super().__init__( vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , low_res_scheduler=lowerCamelCase , scheduler=lowerCamelCase , max_noise_level=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ): torch.manual_seed(0 ) _snake_case = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) _snake_case = add_overlap_rect(lowerCamelCase , lowerCamelCase , image.size ) _snake_case = image.crop(lowerCamelCase ) _snake_case = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] _snake_case = translated_slice_x - (original_image_slice / 2) _snake_case = max(0 , lowerCamelCase ) _snake_case = squeeze_tile(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) _snake_case = to_input.size _snake_case = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) _snake_case = super(lowerCamelCase , self ).__call__(image=lowerCamelCase , **lowerCamelCase ).images[0] _snake_case = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) _snake_case = unsqueeze_tile(lowerCamelCase , lowerCamelCase ) _snake_case = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) _snake_case = [] if x == 0: remove_borders.append("l" ) elif crop_rect[2] == image.size[0]: remove_borders.append("r" ) if y == 0: remove_borders.append("t" ) elif crop_rect[3] == image.size[1]: remove_borders.append("b" ) _snake_case = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=lowerCamelCase ) , mode="L" , ) final_image.paste( lowerCamelCase , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , lowerCamelCase ) @torch.no_grad() def __call__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = 75 , lowerCamelCase = 9.0 , lowerCamelCase = 50 , lowerCamelCase = None , lowerCamelCase = 1 , lowerCamelCase = 0.0 , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = 1 , lowerCamelCase = 128 , lowerCamelCase = 32 , lowerCamelCase = 32 , ): _snake_case = Image.new("RGB" , (image.size[0] * 4, image.size[1] * 4) ) _snake_case = math.ceil(image.size[0] / tile_size ) _snake_case = math.ceil(image.size[1] / tile_size ) _snake_case = tcx * tcy _snake_case = 0 for y in range(lowerCamelCase ): for x in range(lowerCamelCase ): self._process_tile( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , prompt=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , noise_level=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , ) current_count += 1 if callback is not None: callback({"progress": current_count / total_tile_count, "image": final_image} ) return final_image def snake_case_ ( ): '''simple docstring''' _snake_case = "stabilityai/stable-diffusion-x4-upscaler" _snake_case = StableDiffusionTiledUpscalePipeline.from_pretrained(SCREAMING_SNAKE_CASE__ , revision="fp16" , torch_dtype=torch.floataa ) _snake_case = pipe.to("cuda" ) _snake_case = Image.open("../../docs/source/imgs/diffusers_library.jpg" ) def callback(SCREAMING_SNAKE_CASE__ ): print(f'''progress: {obj["progress"]:.4f}''' ) obj["image"].save("diffusers_library_progress.jpg" ) _snake_case = pipe(image=SCREAMING_SNAKE_CASE__ , prompt="Black font, white background, vector" , noise_level=40 , callback=SCREAMING_SNAKE_CASE__ ) final_image.save("diffusers_library.jpg" ) if __name__ == "__main__": main()
672
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __magic_name__ : Optional[int] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"shortest_edge": 256} _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) _snake_case = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
1
'''simple docstring''' import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = tempfile.mkdtemp() _snake_case = 8 # DPR tok _snake_case = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] _snake_case = os.path.join(self.tmpdirname , "dpr_tokenizer" ) os.makedirs(lowerCamelCase , exist_ok=lowerCamelCase ) _snake_case = os.path.join(lowerCamelCase , DPR_VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) # BART tok _snake_case = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] _snake_case = dict(zip(lowerCamelCase , range(len(lowerCamelCase ) ) ) ) _snake_case = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] _snake_case = {"unk_token": "<unk>"} _snake_case = os.path.join(self.tmpdirname , "bart_tokenizer" ) os.makedirs(lowerCamelCase , exist_ok=lowerCamelCase ) _snake_case = os.path.join(lowerCamelCase , BART_VOCAB_FILES_NAMES["vocab_file"] ) _snake_case = os.path.join(lowerCamelCase , BART_VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCamelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowerCamelCase ) ) def UpperCamelCase( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def UpperCamelCase( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , "bart_tokenizer" ) ) def UpperCamelCase( self ): shutil.rmtree(self.tmpdirname ) @require_tokenizers def UpperCamelCase( self ): _snake_case = os.path.join(self.tmpdirname , "rag_tokenizer" ) _snake_case = RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) _snake_case = RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(lowerCamelCase ) rag_tokenizer.save_pretrained(lowerCamelCase ) _snake_case = RagTokenizer.from_pretrained(lowerCamelCase , config=lowerCamelCase ) self.assertIsInstance(new_rag_tokenizer.question_encoder , lowerCamelCase ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , lowerCamelCase ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def UpperCamelCase( self ): _snake_case = RagTokenizer.from_pretrained("facebook/rag-token-nq" ) _snake_case = [ "who got the first nobel prize in physics", "when is the next deadpool movie being released", "which mode is used for short wave broadcast service", "who is the owner of reading football club", "when is the next scandal episode coming out", "when is the last time the philadelphia won the superbowl", "what is the most current adobe flash player version", "how many episodes are there in dragon ball z", "what is the first step in the evolution of the eye", "where is gall bladder situated in human body", "what is the main mineral in lithium batteries", "who is the president of usa right now", "where do the greasers live in the outsiders", "panda is a national animal of which country", "what is the name of manchester united stadium", ] _snake_case = tokenizer(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @slow def UpperCamelCase( self ): _snake_case = RagTokenizer.from_pretrained("facebook/rag-sequence-nq" ) _snake_case = [ "who got the first nobel prize in physics", "when is the next deadpool movie being released", "which mode is used for short wave broadcast service", "who is the owner of reading football club", "when is the next scandal episode coming out", "when is the last time the philadelphia won the superbowl", "what is the most current adobe flash player version", "how many episodes are there in dragon ball z", "what is the first step in the evolution of the eye", "where is gall bladder situated in human body", "what is the main mineral in lithium batteries", "who is the president of usa right now", "where do the greasers live in the outsiders", "panda is a national animal of which country", "what is the name of manchester united stadium", ] _snake_case = tokenizer(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase )
672
'''simple docstring''' import baseaa def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaadecode(SCREAMING_SNAKE_CASE__ ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' __magic_name__ : List[str] = tuple[float, float, float] __magic_name__ : Union[str, Any] = tuple[float, float, float] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = end_pointa[0] - end_pointa[0] _snake_case = end_pointa[1] - end_pointa[1] _snake_case = end_pointa[2] - end_pointa[2] return (x, y, z) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = ab[1] * ac[2] - ab[2] * ac[1] # *i _snake_case = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j _snake_case = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return tuple(round(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for x in vector ) == (0, 0, 0) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 10 ): '''simple docstring''' _snake_case = create_vector(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = create_vector(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return is_zero_vector(get_ad_vectors_cross(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ )
672
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
672
1
'''simple docstring''' import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase( self ): torch.manual_seed(0 ) _snake_case = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def UpperCamelCase( self ): _snake_case = self.dummy_uncond_unet _snake_case = ScoreSdeVeScheduler() _snake_case = ScoreSdeVePipeline(unet=lowerCamelCase , scheduler=lowerCamelCase ) sde_ve.to(lowerCamelCase ) sde_ve.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = torch.manual_seed(0 ) _snake_case = sde_ve(num_inference_steps=2 , output_type="numpy" , generator=lowerCamelCase ).images _snake_case = torch.manual_seed(0 ) _snake_case = sde_ve(num_inference_steps=2 , output_type="numpy" , generator=lowerCamelCase , return_dict=lowerCamelCase )[ 0 ] _snake_case = image[0, -3:, -3:, -1] _snake_case = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _snake_case = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = "google/ncsnpp-church-256" _snake_case = UNetaDModel.from_pretrained(lowerCamelCase ) _snake_case = ScoreSdeVeScheduler.from_pretrained(lowerCamelCase ) _snake_case = ScoreSdeVePipeline(unet=lowerCamelCase , scheduler=lowerCamelCase ) sde_ve.to(lowerCamelCase ) sde_ve.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = torch.manual_seed(0 ) _snake_case = sde_ve(num_inference_steps=10 , output_type="numpy" , generator=lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _snake_case = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
1
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __magic_name__ : Any = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") @dataclass class __SCREAMING_SNAKE_CASE : '''simple docstring''' UpperCAmelCase__ : Optional[str] = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) UpperCAmelCase__ : Optional[str] = field( default=__UpperCamelCase , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) UpperCAmelCase__ : Optional[str] = field( default=__UpperCamelCase , metadata={'''help''': '''The column name of the images in the files.'''} ) UpperCAmelCase__ : Optional[str] = field(default=__UpperCamelCase , metadata={'''help''': '''A folder containing the training data.'''} ) UpperCAmelCase__ : Optional[str] = field(default=__UpperCamelCase , metadata={'''help''': '''A folder containing the validation data.'''} ) UpperCAmelCase__ : Optional[float] = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) UpperCAmelCase__ : Optional[int] = field( default=__UpperCamelCase , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) UpperCAmelCase__ : Optional[int] = field( default=__UpperCamelCase , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def UpperCamelCase( self ): _snake_case = {} if self.train_dir is not None: _snake_case = self.train_dir if self.validation_dir is not None: _snake_case = self.validation_dir _snake_case = data_files if data_files else None @dataclass class __SCREAMING_SNAKE_CASE : '''simple docstring''' UpperCAmelCase__ : str = field( default=__UpperCamelCase , metadata={ '''help''': ( '''The model checkpoint for weights initialization.Don\'t set if you want to train a model from scratch.''' ) } , ) UpperCAmelCase__ : Optional[str] = field( default=__UpperCamelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name_or_path'''} ) UpperCAmelCase__ : Optional[str] = field( default=__UpperCamelCase , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) UpperCAmelCase__ : Optional[str] = field( default=__UpperCamelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from s3'''} ) UpperCAmelCase__ : str = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) UpperCAmelCase__ : str = field(default=__UpperCamelCase , metadata={'''help''': '''Name or path of preprocessor config.'''} ) UpperCAmelCase__ : bool = field( default=__UpperCamelCase , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) UpperCAmelCase__ : float = field( default=0.75 , metadata={'''help''': '''The ratio of the number of masked tokens in the input sequence.'''} ) UpperCAmelCase__ : bool = field( default=__UpperCamelCase , metadata={'''help''': '''Whether or not to train with normalized pixel values as target.'''} ) @dataclass class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : float = field( default=1E-3 , metadata={'''help''': '''Base learning rate: absolute_lr = base_lr * total_batch_size / 256.'''} ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = torch.stack([example["pixel_values"] for example in examples] ) return {"pixel_values": pixel_values} def snake_case_ ( ): '''simple docstring''' _snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _snake_case , _snake_case , _snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _snake_case , _snake_case , _snake_case = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_mae" , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _snake_case = training_args.get_process_log_level() logger.setLevel(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(f'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Initialize our dataset. _snake_case = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. _snake_case = None if "validation" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , SCREAMING_SNAKE_CASE__ ) and data_args.train_val_split > 0.0: _snake_case = ds["train"].train_test_split(data_args.train_val_split ) _snake_case = split["train"] _snake_case = split["test"] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _snake_case = { "cache_dir": model_args.cache_dir, "revision": model_args.model_revision, "use_auth_token": True if model_args.use_auth_token else None, } if model_args.config_name: _snake_case = ViTMAEConfig.from_pretrained(model_args.config_name , **SCREAMING_SNAKE_CASE__ ) elif model_args.model_name_or_path: _snake_case = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **SCREAMING_SNAKE_CASE__ ) else: _snake_case = ViTMAEConfig() logger.warning("You are instantiating a new config instance from scratch." ) if model_args.config_overrides is not None: logger.info(f'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(f'''New config: {config}''' ) # adapt config config.update( { "mask_ratio": model_args.mask_ratio, "norm_pix_loss": model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: _snake_case = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **SCREAMING_SNAKE_CASE__ ) elif model_args.model_name_or_path: _snake_case = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **SCREAMING_SNAKE_CASE__ ) else: _snake_case = ViTImageProcessor() # create model if model_args.model_name_or_path: _snake_case = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("Training new model from scratch" ) _snake_case = ViTMAEForPreTraining(SCREAMING_SNAKE_CASE__ ) if training_args.do_train: _snake_case = ds["train"].column_names else: _snake_case = ds["validation"].column_names if data_args.image_column_name is not None: _snake_case = data_args.image_column_name elif "image" in column_names: _snake_case = "image" elif "img" in column_names: _snake_case = "img" else: _snake_case = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: _snake_case = image_processor.size["shortest_edge"] else: _snake_case = (image_processor.size["height"], image_processor.size["width"]) _snake_case = Compose( [ Lambda(lambda SCREAMING_SNAKE_CASE__ : img.convert("RGB" ) if img.mode != "RGB" else img ), RandomResizedCrop(SCREAMING_SNAKE_CASE__ , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(SCREAMING_SNAKE_CASE__ ): _snake_case = [transforms(SCREAMING_SNAKE_CASE__ ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _snake_case = ds["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(SCREAMING_SNAKE_CASE__ ) if training_args.do_eval: if "validation" not in ds: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _snake_case = ( ds["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(SCREAMING_SNAKE_CASE__ ) # Compute absolute learning rate _snake_case = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: _snake_case = training_args.base_learning_rate * total_train_batch_size / 2_56 # Initialize our trainer _snake_case = Trainer( model=SCREAMING_SNAKE_CASE__ , args=SCREAMING_SNAKE_CASE__ , train_dataset=ds["train"] if training_args.do_train else None , eval_dataset=ds["validation"] if training_args.do_eval else None , tokenizer=SCREAMING_SNAKE_CASE__ , data_collator=SCREAMING_SNAKE_CASE__ , ) # Training if training_args.do_train: _snake_case = None if training_args.resume_from_checkpoint is not None: _snake_case = training_args.resume_from_checkpoint elif last_checkpoint is not None: _snake_case = last_checkpoint _snake_case = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE__ ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _snake_case = trainer.evaluate() trainer.log_metrics("eval" , SCREAMING_SNAKE_CASE__ ) trainer.save_metrics("eval" , SCREAMING_SNAKE_CASE__ ) # Write model card and (optionally) push to hub _snake_case = { "tasks": "masked-auto-encoding", "dataset": data_args.dataset_name, "tags": ["masked-auto-encoding"], } if training_args.push_to_hub: trainer.push_to_hub(**SCREAMING_SNAKE_CASE__ ) else: trainer.create_model_card(**SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' main() if __name__ == "__main__": main()
672
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
1
'''simple docstring''' import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa __magic_name__ : Tuple = logging.getLogger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = '''summarization''' UpperCAmelCase__ : Optional[Any] = ['''loss'''] UpperCAmelCase__ : Tuple = ROUGE_KEYS UpperCAmelCase__ : Union[str, Any] = '''rouge2''' def __init__( self , lowerCamelCase , **lowerCamelCase ): if hparams.sortish_sampler and hparams.gpus > 1: _snake_case = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError("Dynamic Batch size does not work for multi-gpu training" ) if hparams.sortish_sampler: raise ValueError("--sortish_sampler and --max_tokens_per_batch may not be used simultaneously" ) super().__init__(lowerCamelCase , num_labels=lowerCamelCase , mode=self.mode , **lowerCamelCase ) use_task_specific_params(self.model , "summarization" ) save_git_info(self.hparams.output_dir ) _snake_case = Path(self.output_dir ) / "metrics.json" _snake_case = Path(self.output_dir ) / "hparams.pkl" pickle_save(self.hparams , self.hparams_save_path ) _snake_case = 0 _snake_case = defaultdict(lowerCamelCase ) _snake_case = self.config.model_type _snake_case = self.config.tgt_vocab_size if self.model_type == "fsmt" else self.config.vocab_size _snake_case = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } _snake_case = { "train": self.hparams.n_train, "val": self.hparams.n_val, "test": self.hparams.n_test, } _snake_case = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} _snake_case = { "train": self.hparams.max_target_length, "val": self.hparams.val_max_target_length, "test": self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], F'''target_lens: {self.target_lens}''' assert self.target_lens["train"] <= self.target_lens["test"], F'''target_lens: {self.target_lens}''' if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) _snake_case = get_git_info()["repo_sha"] _snake_case = hparams.num_workers _snake_case = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , lowerCamelCase ): _snake_case = self.tokenizer.lang_code_to_id[hparams.tgt_lang] _snake_case = self.decoder_start_token_id _snake_case = ( SeqaSeqDataset if hasattr(self.tokenizer , "prepare_seq2seq_batch" ) else LegacySeqaSeqDataset ) _snake_case = False _snake_case = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: _snake_case = self.hparams.eval_max_gen_length else: _snake_case = self.model.config.max_length _snake_case = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def UpperCamelCase( self , lowerCamelCase ): _snake_case = { k: self.tokenizer.batch_decode(v.tolist() ) if "mask" not in k else v.shape for k, v in batch.items() } save_json(lowerCamelCase , Path(self.output_dir ) / "text_batch.json" ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / "tok_batch.json" ) _snake_case = True return readable_batch def UpperCamelCase( self , lowerCamelCase , **lowerCamelCase ): return self.model(lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = self.tokenizer.batch_decode( lowerCamelCase , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase ) return lmap(str.strip , lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = self.tokenizer.pad_token_id _snake_case , _snake_case = batch["input_ids"], batch["attention_mask"] _snake_case = batch["labels"] if isinstance(self.model , lowerCamelCase ): _snake_case = self.model._shift_right(lowerCamelCase ) else: _snake_case = shift_tokens_right(lowerCamelCase , lowerCamelCase ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero _snake_case = decoder_input_ids self.save_readable_batch(lowerCamelCase ) _snake_case = self(lowerCamelCase , attention_mask=lowerCamelCase , decoder_input_ids=lowerCamelCase , use_cache=lowerCamelCase ) _snake_case = outputs["logits"] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id _snake_case = nn.CrossEntropyLoss(ignore_index=lowerCamelCase ) assert lm_logits.shape[-1] == self.vocab_size _snake_case = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: _snake_case = nn.functional.log_softmax(lowerCamelCase , dim=-1 ) _snake_case , _snake_case = label_smoothed_nll_loss( lowerCamelCase , lowerCamelCase , self.hparams.label_smoothing , ignore_index=lowerCamelCase ) return (loss,) @property def UpperCamelCase( self ): return self.tokenizer.pad_token_id def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self._step(lowerCamelCase ) _snake_case = dict(zip(self.loss_names , lowerCamelCase ) ) # tokens per batch _snake_case = batch["input_ids"].ne(self.pad ).sum() + batch["labels"].ne(self.pad ).sum() _snake_case = batch["input_ids"].shape[0] _snake_case = batch["input_ids"].eq(self.pad ).sum() _snake_case = batch["input_ids"].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return self._generative_step(lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase="val" ): self.step_count += 1 _snake_case = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} _snake_case = losses["loss"] _snake_case = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ["gen_time", "gen_len"] } _snake_case = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) _snake_case = torch.tensor(lowerCamelCase ).type_as(lowerCamelCase ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(lowerCamelCase ) _snake_case = {F'''{prefix}_avg_{k}''': x for k, x in losses.items()} _snake_case = self.step_count self.metrics[prefix].append(lowerCamelCase ) # callback writes this to self.metrics_save_path _snake_case = flatten_list([x["preds"] for x in outputs] ) return { "log": all_metrics, "preds": preds, F'''{prefix}_loss''': loss, F'''{prefix}_{self.val_metric}''': metric_tensor, } def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return calculate_rouge(lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') _snake_case = self.model.generate( batch["input_ids"] , attention_mask=batch["attention_mask"] , use_cache=lowerCamelCase , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) _snake_case = (time.time() - ta) / batch["input_ids"].shape[0] _snake_case = self.ids_to_clean_text(lowerCamelCase ) _snake_case = self.ids_to_clean_text(batch["labels"] ) _snake_case = self._step(lowerCamelCase ) _snake_case = dict(zip(self.loss_names , lowerCamelCase ) ) _snake_case = self.calc_generative_metrics(lowerCamelCase , lowerCamelCase ) _snake_case = np.mean(lmap(lowerCamelCase , lowerCamelCase ) ) base_metrics.update(gen_time=lowerCamelCase , gen_len=lowerCamelCase , preds=lowerCamelCase , target=lowerCamelCase , **lowerCamelCase ) return base_metrics def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return self._generative_step(lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase ): return self.validation_epoch_end(lowerCamelCase , prefix="test" ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = self.n_obs[type_path] _snake_case = self.target_lens[type_path] _snake_case = self.dataset_class( self.tokenizer , type_path=lowerCamelCase , n_obs=lowerCamelCase , max_target_length=lowerCamelCase , **self.dataset_kwargs , ) return dataset def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = False ): _snake_case = self.get_dataset(lowerCamelCase ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": _snake_case = dataset.make_sortish_sampler(lowerCamelCase , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCamelCase , batch_size=lowerCamelCase , collate_fn=dataset.collate_fn , shuffle=lowerCamelCase , num_workers=self.num_workers , sampler=lowerCamelCase , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": _snake_case = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCamelCase , batch_sampler=lowerCamelCase , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( lowerCamelCase , batch_size=lowerCamelCase , collate_fn=dataset.collate_fn , shuffle=lowerCamelCase , num_workers=self.num_workers , sampler=lowerCamelCase , ) def UpperCamelCase( self ): _snake_case = self.get_dataloader("train" , batch_size=self.hparams.train_batch_size , shuffle=lowerCamelCase ) return dataloader def UpperCamelCase( self ): return self.get_dataloader("val" , batch_size=self.hparams.eval_batch_size ) def UpperCamelCase( self ): return self.get_dataloader("test" , batch_size=self.hparams.eval_batch_size ) @staticmethod def UpperCamelCase( lowerCamelCase , lowerCamelCase ): BaseTransformer.add_model_specific_args(lowerCamelCase , lowerCamelCase ) add_generic_args(lowerCamelCase , lowerCamelCase ) parser.add_argument( "--max_source_length" , default=1_024 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--max_target_length" , default=56 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--val_max_target_length" , default=142 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--test_max_target_length" , default=142 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument("--freeze_encoder" , action="store_true" ) parser.add_argument("--freeze_embeds" , action="store_true" ) parser.add_argument("--sortish_sampler" , action="store_true" , default=lowerCamelCase ) parser.add_argument("--overwrite_output_dir" , action="store_true" , default=lowerCamelCase ) parser.add_argument("--max_tokens_per_batch" , type=lowerCamelCase , default=lowerCamelCase ) parser.add_argument("--logger_name" , type=lowerCamelCase , choices=["default", "wandb", "wandb_shared"] , default="default" ) parser.add_argument("--n_train" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_val" , type=lowerCamelCase , default=500 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_test" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument( "--task" , type=lowerCamelCase , default="summarization" , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--label_smoothing" , type=lowerCamelCase , default=0.0 , required=lowerCamelCase ) parser.add_argument("--src_lang" , type=lowerCamelCase , default="" , required=lowerCamelCase ) parser.add_argument("--tgt_lang" , type=lowerCamelCase , default="" , required=lowerCamelCase ) parser.add_argument("--eval_beams" , type=lowerCamelCase , default=lowerCamelCase , required=lowerCamelCase ) parser.add_argument( "--val_metric" , type=lowerCamelCase , default=lowerCamelCase , required=lowerCamelCase , choices=["bleu", "rouge2", "loss", None] ) parser.add_argument("--eval_max_gen_length" , type=lowerCamelCase , default=lowerCamelCase , help="never generate more than n tokens" ) parser.add_argument("--save_top_k" , type=lowerCamelCase , default=1 , required=lowerCamelCase , help="How many checkpoints to save" ) parser.add_argument( "--early_stopping_patience" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help=( "-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So" " val_check_interval will effect it." ) , ) return parser class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : str = '''translation''' UpperCAmelCase__ : int = ['''loss'''] UpperCAmelCase__ : Optional[int] = ['''bleu'''] UpperCAmelCase__ : Tuple = '''bleu''' def __init__( self , lowerCamelCase , **lowerCamelCase ): super().__init__(lowerCamelCase , **lowerCamelCase ) _snake_case = hparams.src_lang _snake_case = hparams.tgt_lang def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return calculate_bleu(lowerCamelCase , lowerCamelCase ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None ): '''simple docstring''' Path(args.output_dir ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) check_output_dir(SCREAMING_SNAKE_CASE__ , expected_items=3 ) if model is None: if "summarization" in args.task: _snake_case = SummarizationModule(SCREAMING_SNAKE_CASE__ ) else: _snake_case = TranslationModule(SCREAMING_SNAKE_CASE__ ) _snake_case = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith("/tmp" ) or str(args.output_dir ).startswith("/var" ) ): _snake_case = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger _snake_case = os.environ.get("WANDB_PROJECT" , SCREAMING_SNAKE_CASE__ ) _snake_case = WandbLogger(name=model.output_dir.name , project=SCREAMING_SNAKE_CASE__ ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger _snake_case = WandbLogger(name=model.output_dir.name , project=f'''hf_{dataset}''' ) if args.early_stopping_patience >= 0: _snake_case = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: _snake_case = False _snake_case = args.val_metric == "loss" _snake_case = generic_train( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , SCREAMING_SNAKE_CASE__ ) , early_stopping_callback=SCREAMING_SNAKE_CASE__ , logger=SCREAMING_SNAKE_CASE__ , ) pickle_save(model.hparams , model.output_dir / "hparams.pkl" ) if not args.do_predict: return model _snake_case = "" _snake_case = sorted(glob.glob(os.path.join(args.output_dir , "*.ckpt" ) , recursive=SCREAMING_SNAKE_CASE__ ) ) if checkpoints: _snake_case = checkpoints[-1] _snake_case = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": __magic_name__ : int = argparse.ArgumentParser() __magic_name__ : str = pl.Trainer.add_argparse_args(parser) __magic_name__ : Optional[int] = SummarizationModule.add_model_specific_args(parser, os.getcwd()) __magic_name__ : Optional[Any] = parser.parse_args() main(args)
672
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = text_path elif issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [text_path] _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=("train",) ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for split in splits: _snake_case = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader({"train": text_path} , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader({"train": text_path} , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if split: _snake_case = {split: text_path} else: _snake_case = "train" _snake_case = {"train": text_path, "test": text_path} _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
672
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __magic_name__ : Any = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = b.T _snake_case = np.sum(np.square(SCREAMING_SNAKE_CASE__ ) , axis=1 ) _snake_case = np.sum(np.square(SCREAMING_SNAKE_CASE__ ) , axis=0 ) _snake_case = np.matmul(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = aa[:, None] - 2 * ab + ba[None, :] return d def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = x.reshape(-1 , 3 ) _snake_case = squared_euclidean_distance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return np.argmin(SCREAMING_SNAKE_CASE__ , axis=1 ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = ['''pixel_values'''] def __init__( self , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"height": 256, "width": 256} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = np.array(lowerCamelCase ) if clusters is not None else None _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_normalize _snake_case = do_color_quantize def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F'''Size dictionary must contain both height and width keys. Got {size.keys()}''' ) return resize( lowerCamelCase , size=(size["height"], size["width"]) , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , ): _snake_case = rescale(image=lowerCamelCase , scale=1 / 127.5 , data_format=lowerCamelCase ) _snake_case = image - 1 return image def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = do_color_quantize if do_color_quantize is not None else self.do_color_quantize _snake_case = clusters if clusters is not None else self.clusters _snake_case = np.array(lowerCamelCase ) _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_color_quantize and clusters is None: raise ValueError("Clusters must be specified if do_color_quantize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase ) for image in images] if do_color_quantize: _snake_case = [to_channel_dimension_format(lowerCamelCase , ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) _snake_case = np.array(lowerCamelCase ) _snake_case = color_quantize(lowerCamelCase , lowerCamelCase ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) _snake_case = images.shape[0] _snake_case = images.reshape(lowerCamelCase , -1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. _snake_case = list(lowerCamelCase ) else: _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"input_ids": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ : Any = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __magic_name__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[Any] = logging.get_logger(__name__) __magic_name__ : Tuple = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/config.json""" # See all FNet models at https://huggingface.co/models?filter=fnet } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = '''fnet''' def __init__( self , lowerCamelCase=32_000 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=4 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=False , lowerCamelCase=512 , lowerCamelCase=3 , lowerCamelCase=1 , lowerCamelCase=2 , **lowerCamelCase , ): super().__init__(pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , **lowerCamelCase ) _snake_case = vocab_size _snake_case = max_position_embeddings _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = initializer_range _snake_case = type_vocab_size _snake_case = layer_norm_eps _snake_case = use_tpu_fourier_optimizations _snake_case = tpu_short_seq_length
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE__ , 2 ) - a def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 2 * x def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2.0 while start <= a: _snake_case = math.pow(SCREAMING_SNAKE_CASE__ , 2 ) return start def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 99_99 , SCREAMING_SNAKE_CASE__ = 0.00000000000001 ): '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _snake_case = get_initial_point(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = value _snake_case = value - fx(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) / fx_derivative(SCREAMING_SNAKE_CASE__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
672
1
'''simple docstring''' from itertools import product def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = sides_number _snake_case = max_face_number * dice_number _snake_case = [0] * (max_total + 1) _snake_case = 1 _snake_case = range(SCREAMING_SNAKE_CASE__ , max_face_number + 1 ) for dice_numbers in product(SCREAMING_SNAKE_CASE__ , repeat=SCREAMING_SNAKE_CASE__ ): _snake_case = sum(SCREAMING_SNAKE_CASE__ ) totals_frequencies[total] += 1 return totals_frequencies def snake_case_ ( ): '''simple docstring''' _snake_case = total_frequency_distribution( sides_number=4 , dice_number=9 ) _snake_case = total_frequency_distribution( sides_number=6 , dice_number=6 ) _snake_case = 0 _snake_case = 9 _snake_case = 4 * 9 _snake_case = 6 for peter_total in range(SCREAMING_SNAKE_CASE__ , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) _snake_case = (4**9) * (6**6) _snake_case = peter_wins_count / total_games_number _snake_case = round(SCREAMING_SNAKE_CASE__ , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(F'{solution() = }')
672
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[int] = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = '''git_vision_model''' def __init__( self , lowerCamelCase=768 , lowerCamelCase=3_072 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=16 , lowerCamelCase="quick_gelu" , lowerCamelCase=1e-5 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = num_channels _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = '''git''' def __init__( self , lowerCamelCase=None , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=6 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1_024 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=101 , lowerCamelCase=102 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , pad_token_id=lowerCamelCase , **lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) _snake_case = GitVisionConfig(**lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = num_image_with_embedding _snake_case = bos_token_id _snake_case = eos_token_id def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.__class__.model_type return output
672
1
'''simple docstring''' import math import random from typing import Any from .hill_climbing import SearchProblem def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = True , SCREAMING_SNAKE_CASE__ = math.inf , SCREAMING_SNAKE_CASE__ = -math.inf , SCREAMING_SNAKE_CASE__ = math.inf , SCREAMING_SNAKE_CASE__ = -math.inf , SCREAMING_SNAKE_CASE__ = False , SCREAMING_SNAKE_CASE__ = 1_00 , SCREAMING_SNAKE_CASE__ = 0.01 , SCREAMING_SNAKE_CASE__ = 1 , ): '''simple docstring''' _snake_case = False _snake_case = search_prob _snake_case = start_temperate _snake_case = [] _snake_case = 0 _snake_case = None while not search_end: _snake_case = current_state.score() if best_state is None or current_score > best_state.score(): _snake_case = current_state scores.append(SCREAMING_SNAKE_CASE__ ) iterations += 1 _snake_case = None _snake_case = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to _snake_case = random.randint(0 , len(SCREAMING_SNAKE_CASE__ ) - 1 ) # picking a random neighbor _snake_case = neighbors.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: _snake_case = change * -1 # in case we are finding minimum if change > 0: # improves the solution _snake_case = picked_neighbor else: _snake_case = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability _snake_case = picked_neighbor _snake_case = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor _snake_case = True else: _snake_case = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) plt.xlabel("Iterations" ) plt.ylabel("Function values" ) plt.show() return best_state if __name__ == "__main__": def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) __magic_name__ : Dict = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __magic_name__ : List[str] = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( """The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) # starting the problem with initial coordinates (12, 47) __magic_name__ : Tuple = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __magic_name__ : str = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( """The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return (3 * x**2) - (6 * y) __magic_name__ : Tuple = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __magic_name__ : List[Any] = simulated_annealing(prob, find_max=False, visualization=True) print( """The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ F'{local_min.score()}' ) __magic_name__ : List[str] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __magic_name__ : Any = simulated_annealing(prob, find_max=True, visualization=True) print( """The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ F'{local_min.score()}' )
672
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __magic_name__ : Dict = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , np.ndarray ): return list(tensor.shape ) _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) if tensor.shape == tf.TensorShape(SCREAMING_SNAKE_CASE__ ): return dynamic _snake_case = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(SCREAMING_SNAKE_CASE__ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise NotImplementedError("Only 1D weight and bias tensors are supported for now, with only a single axis." ) # Get mean and variance on the axis to be normalized _snake_case , _snake_case = tf.nn.moments(SCREAMING_SNAKE_CASE__ , axes=[axis] , keepdims=SCREAMING_SNAKE_CASE__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis _snake_case = [1] * inputs.shape.rank _snake_case = shape_list(SCREAMING_SNAKE_CASE__ )[axis] _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Compute layer normalization using the batch_normalization # function. _snake_case = tf.nn.batch_normalization( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , offset=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , variance_epsilon=SCREAMING_SNAKE_CASE__ , ) return outputs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) _snake_case = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) _snake_case = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ): _snake_case = tf.convert_to_tensor(SCREAMING_SNAKE_CASE__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: _snake_case = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: _snake_case = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) _snake_case = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( SCREAMING_SNAKE_CASE__ , tf.cast(SCREAMING_SNAKE_CASE__ , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(SCREAMING_SNAKE_CASE__ )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. _snake_case = [x for x in data if len(SCREAMING_SNAKE_CASE__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( "The following attributes cannot be saved to HDF5 file because " f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) _snake_case = np.asarray(SCREAMING_SNAKE_CASE__ ) _snake_case = 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(SCREAMING_SNAKE_CASE__ ): _snake_case = chunk_data else: _snake_case = data def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if name in group.attrs: _snake_case = [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs[name]] else: _snake_case = [] _snake_case = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs["%s%d" % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _expand_single_ad_tensor(SCREAMING_SNAKE_CASE__ ): if isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(SCREAMING_SNAKE_CASE__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , SCREAMING_SNAKE_CASE__ )
672
1
'''simple docstring''' import cmath import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = math.radians(SCREAMING_SNAKE_CASE__ ) _snake_case = math.radians(SCREAMING_SNAKE_CASE__ ) # Convert voltage and current to rectangular form _snake_case = cmath.rect(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = cmath.rect(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
1
'''simple docstring''' import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=64 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=[1, 16, 4, 4] , lowerCamelCase=None , ): _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = patch_size _snake_case = num_channels _snake_case = is_training _snake_case = use_labels _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = scope _snake_case = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size _snake_case = (self.image_size // 32) ** 2 _snake_case = num_patches + 1 def UpperCamelCase( self ): _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase( self ): _snake_case = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [4, 8, 16, 32], "num_groups": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = ViTHybridModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = self.type_sequence_label_size _snake_case = ViTHybridForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase( self ): _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () UpperCAmelCase__ : Optional[int] = ( {'''feature-extraction''': ViTHybridModel, '''image-classification''': ViTHybridForImageClassification} if is_torch_available() else {} ) UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : int = False def UpperCamelCase( self ): _snake_case = ViTHybridModelTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def UpperCamelCase( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(lowerCamelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(lowerCamelCase ) for model_class in self.all_model_classes: _snake_case = model_class(config=lowerCamelCase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": _snake_case = [F'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @slow def UpperCamelCase( self ): for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = ViTHybridModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' _snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase( self ): return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase( self ): _snake_case = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowerCamelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): _snake_case = model(**lowerCamelCase ) # verify the logits _snake_case = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) _snake_case = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate def UpperCamelCase( self ): _snake_case = ViTHybridImageProcessor.from_pretrained("google/vit-hybrid-base-bit-384" ) _snake_case = ViTHybridForImageClassification.from_pretrained("google/vit-hybrid-base-bit-384" , device_map="auto" ) _snake_case = prepare_img() _snake_case = image_processor(images=lowerCamelCase , return_tensors="pt" ) _snake_case = model(**lowerCamelCase ) _snake_case = outputs.logits # model predicts one of the 1000 ImageNet classes _snake_case = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , "tabby, tabby cat" )
672
'''simple docstring''' from torch import nn def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'''Unsupported activation function: {act_fn}''' )
672
1
'''simple docstring''' import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=10 ): '''simple docstring''' _snake_case = [] for _ in range(SCREAMING_SNAKE_CASE__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=10 ): '''simple docstring''' _snake_case = [] for step in range(SCREAMING_SNAKE_CASE__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _snake_case = os.path.join(SCREAMING_SNAKE_CASE__ , "schedule.bin" ) torch.save(scheduler.state_dict() , SCREAMING_SNAKE_CASE__ ) _snake_case = torch.load(SCREAMING_SNAKE_CASE__ ) scheduler.load_state_dict(SCREAMING_SNAKE_CASE__ ) return lrs @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for a, b in zip(lowerCamelCase , lowerCamelCase ): self.assertAlmostEqual(lowerCamelCase , lowerCamelCase , delta=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowerCamelCase ) _snake_case = torch.tensor([0.4, 0.2, -0.5] ) _snake_case = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _snake_case = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(100 ): _snake_case = criterion(lowerCamelCase , lowerCamelCase ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def UpperCamelCase( self ): _snake_case = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowerCamelCase ) _snake_case = torch.tensor([0.4, 0.2, -0.5] ) _snake_case = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _snake_case = Adafactor( params=[w] , lr=1e-2 , eps=(1e-30, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=lowerCamelCase , weight_decay=0.0 , relative_step=lowerCamelCase , scale_parameter=lowerCamelCase , warmup_init=lowerCamelCase , ) for _ in range(1_000 ): _snake_case = criterion(lowerCamelCase , lowerCamelCase ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : str = nn.Linear(50 , 50 ) if is_torch_available() else None UpperCAmelCase__ : List[str] = AdamW(m.parameters() , lr=10.0 ) if is_torch_available() else None UpperCAmelCase__ : List[Any] = 10 def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None ): self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for a, b in zip(lowerCamelCase , lowerCamelCase ): self.assertAlmostEqual(lowerCamelCase , lowerCamelCase , delta=lowerCamelCase , msg=lowerCamelCase ) def UpperCamelCase( self ): _snake_case = {"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _snake_case = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1e-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): _snake_case , _snake_case = data _snake_case = scheduler_func(self.optimizer , **lowerCamelCase ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _snake_case = unwrap_schedule(lowerCamelCase , self.num_steps ) self.assertListAlmostEqual( lowerCamelCase , lowerCamelCase , tol=1e-2 , msg=F'''failed for {scheduler_func} in normal scheduler''' , ) _snake_case = scheduler_func(self.optimizer , **lowerCamelCase ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(lowerCamelCase ) # wrap to test picklability of the schedule _snake_case = unwrap_and_save_reload_schedule(lowerCamelCase , self.num_steps ) self.assertListEqual(lowerCamelCase , lowerCamelCase , msg=F'''failed for {scheduler_func} in save and reload''' ) class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase ): _snake_case = fn def __call__( self , *lowerCamelCase , **lowerCamelCase ): return self.fn(*lowerCamelCase , **lowerCamelCase ) @classmethod def UpperCamelCase( self , lowerCamelCase ): _snake_case = list(map(self , scheduler.lr_lambdas ) )
672
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __magic_name__ : Tuple = 0 __magic_name__ : Dict = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __magic_name__ : str = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __magic_name__ : Dict = tuple[int, int] class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = pos_x _snake_case = pos_y _snake_case = (pos_y, pos_x) _snake_case = goal_x _snake_case = goal_y _snake_case = g_cost _snake_case = parent _snake_case = self.calculate_heuristic() _snake_case = self.g_cost + self.h_cost def UpperCamelCase( self ): _snake_case = self.pos_x - self.goal_x _snake_case = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCamelCase ) + abs(lowerCamelCase ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowerCamelCase ): return self.f_cost < other.f_cost class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCamelCase ) _snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , lowerCamelCase ) _snake_case = [self.start] _snake_case = [] _snake_case = False def UpperCamelCase( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _snake_case = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCamelCase ) self.closed_nodes.append(lowerCamelCase ) _snake_case = self.get_successors(lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = self.open_nodes.pop(self.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCamelCase ) else: self.open_nodes.append(lowerCamelCase ) return [self.start.pos] def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for action in delta: _snake_case = parent.pos_x + action[1] _snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCamelCase , lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCamelCase , ) ) return successors def UpperCamelCase( self , lowerCamelCase ): _snake_case = node _snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _snake_case = current_node.parent path.reverse() return path class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = False def UpperCamelCase( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _snake_case = self.fwd_astar.open_nodes.pop(0 ) _snake_case = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCamelCase , lowerCamelCase ) self.fwd_astar.closed_nodes.append(lowerCamelCase ) self.bwd_astar.closed_nodes.append(lowerCamelCase ) _snake_case = current_bwd_node _snake_case = current_fwd_node _snake_case = { self.fwd_astar: self.fwd_astar.get_successors(lowerCamelCase ), self.bwd_astar: self.bwd_astar.get_successors(lowerCamelCase ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = astar.open_nodes.pop( astar.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCamelCase ) else: astar.open_nodes.append(lowerCamelCase ) return [self.fwd_astar.start.pos] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self.fwd_astar.retrace_path(lowerCamelCase ) _snake_case = self.bwd_astar.retrace_path(lowerCamelCase ) bwd_path.pop() bwd_path.reverse() _snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __magic_name__ : Optional[int] = (0, 0) __magic_name__ : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __magic_name__ : Any = time.time() __magic_name__ : Optional[int] = AStar(init, goal) __magic_name__ : str = a_star.search() __magic_name__ : List[Any] = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') __magic_name__ : List[str] = time.time() __magic_name__ : Optional[Any] = BidirectionalAStar(init, goal) __magic_name__ : Optional[int] = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
672
1
'''simple docstring''' import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase = "cpu" , lowerCamelCase = "openai/clip-vit-large-patch14" ): _snake_case = device _snake_case = CLIPTokenizerFast.from_pretrained(lowerCamelCase ) _snake_case = [0.48145466, 0.4578275, 0.40821073] _snake_case = [0.26862954, 0.26130258, 0.27577711] _snake_case = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _snake_case = torchvision.transforms.Resize(224 ) _snake_case = torchvision.transforms.CenterCrop(224 ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = self.resize(lowerCamelCase ) _snake_case = self.center_crop(lowerCamelCase ) _snake_case = self.normalize(lowerCamelCase ) return images def __call__( self , lowerCamelCase=None , lowerCamelCase=None , **lowerCamelCase ): _snake_case = self.tokenizer(text=lowerCamelCase , **lowerCamelCase ) _snake_case = self.preprocess_img(lowerCamelCase ) _snake_case = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class __SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase=10 , lowerCamelCase=0.01 , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase="image" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=False , ): super().__init__() _snake_case = None _snake_case = device if device else get_device() if vqgan: _snake_case = vqgan else: _snake_case = load_vqgan(self.device , conf_path=lowerCamelCase , ckpt_path=lowerCamelCase ) self.vqgan.eval() if clip: _snake_case = clip else: _snake_case = CLIPModel.from_pretrained("openai/clip-vit-base-patch32" ) self.clip.to(self.device ) _snake_case = ProcessorGradientFlow(device=self.device ) _snake_case = iterations _snake_case = lr _snake_case = log _snake_case = make_grid _snake_case = return_val _snake_case = quantize _snake_case = self.vqgan.decoder.z_shape def UpperCamelCase( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=5 , lowerCamelCase=True ): _snake_case = [] if output_path is None: _snake_case = "./animation.gif" if input_path is None: _snake_case = self.save_path _snake_case = sorted(glob(input_path + "/*" ) ) if not len(lowerCamelCase ): raise ValueError( "No images found in save path, aborting (did you pass save_intermediate=True to the generate" " function?)" ) if len(lowerCamelCase ) == 1: print("Only one image found in save path, (did you pass save_intermediate=True to the generate function?)" ) _snake_case = total_duration / len(lowerCamelCase ) _snake_case = [frame_duration] * len(lowerCamelCase ) if extend_frames: _snake_case = 1.5 _snake_case = 3 for file_name in paths: if file_name.endswith(".png" ): images.append(imageio.imread(lowerCamelCase ) ) imageio.mimsave(lowerCamelCase , lowerCamelCase , duration=lowerCamelCase ) print(F'''gif saved to {output_path}''' ) def UpperCamelCase( self , lowerCamelCase=None , lowerCamelCase=None ): if not (path or img): raise ValueError("Input either path or tensor" ) if img is not None: raise NotImplementedError _snake_case = preprocess(Image.open(lowerCamelCase ) , target_image_size=256 ).to(self.device ) _snake_case = preprocess_vqgan(lowerCamelCase ) _snake_case , *_snake_case = self.vqgan.encode(lowerCamelCase ) return z def UpperCamelCase( self , lowerCamelCase ): _snake_case = self.latent.detach().requires_grad_() _snake_case = base_latent + transform_vector if self.quantize: _snake_case , *_snake_case = self.vqgan.quantize(lowerCamelCase ) else: _snake_case = trans_latent return self.vqgan.decode(lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=None ): _snake_case = self.clip_preprocessor(text=lowerCamelCase , images=lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ) _snake_case = self.clip(**lowerCamelCase ) _snake_case = clip_outputs.logits_per_image if weights is not None: _snake_case = similarity_logits * weights return similarity_logits.sum() def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = self._get_clip_similarity(pos_prompts["prompts"] , lowerCamelCase , weights=(1 / pos_prompts["weights"]) ) if neg_prompts: _snake_case = self._get_clip_similarity(neg_prompts["prompts"] , lowerCamelCase , weights=neg_prompts["weights"] ) else: _snake_case = torch.tensor([1] , device=self.device ) _snake_case = -torch.log(lowerCamelCase ) + torch.log(lowerCamelCase ) return loss def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = torch.randn_like(self.latent , requires_grad=lowerCamelCase , device=self.device ) _snake_case = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _snake_case = self._add_vector(lowerCamelCase ) _snake_case = loop_post_process(lowerCamelCase ) _snake_case = self._get_CLIP_loss(lowerCamelCase , lowerCamelCase , lowerCamelCase ) print("CLIP loss" , lowerCamelCase ) if self.log: wandb.log({"CLIP Loss": clip_loss} ) clip_loss.backward(retain_graph=lowerCamelCase ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): wandb.init(reinit=lowerCamelCase , project="face-editor" ) wandb.config.update({"Positive Prompts": positive_prompts} ) wandb.config.update({"Negative Prompts": negative_prompts} ) wandb.config.update({"lr": self.lr, "iterations": self.iterations} ) if image_path: _snake_case = Image.open(lowerCamelCase ) _snake_case = image.resize((256, 256) ) wandb.log("Original Image" , wandb.Image(lowerCamelCase ) ) def UpperCamelCase( self , lowerCamelCase ): if not prompts: return [] _snake_case = [] _snake_case = [] if isinstance(lowerCamelCase , lowerCamelCase ): _snake_case = [prompt.strip() for prompt in prompts.split("|" )] for prompt in prompts: if isinstance(lowerCamelCase , (tuple, list) ): _snake_case = prompt[0] _snake_case = float(prompt[1] ) elif ":" in prompt: _snake_case , _snake_case = prompt.split(":" ) _snake_case = float(lowerCamelCase ) else: _snake_case = prompt _snake_case = 1.0 processed_prompts.append(lowerCamelCase ) weights.append(lowerCamelCase ) return { "prompts": processed_prompts, "weights": torch.tensor(lowerCamelCase , device=self.device ), } def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=None , ): if image_path: _snake_case = self._get_latent(lowerCamelCase ) else: _snake_case = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(lowerCamelCase , lowerCamelCase , lowerCamelCase ) assert pos_prompts, "You must provide at least one positive prompt." _snake_case = self.process_prompts(lowerCamelCase ) _snake_case = self.process_prompts(lowerCamelCase ) if save_final and save_path is None: _snake_case = os.path.join("./outputs/" , "_".join(pos_prompts["prompts"] ) ) if not os.path.exists(lowerCamelCase ): os.makedirs(lowerCamelCase ) else: _snake_case = save_path + "_" + get_timestamp() os.makedirs(lowerCamelCase ) _snake_case = save_path _snake_case = self.vqgan.decode(self.latent )[0] if show_intermediate: print("Original Image" ) show_pil(custom_to_pil(lowerCamelCase ) ) _snake_case = loop_post_process(lowerCamelCase ) for iter, transformed_img in enumerate(self._optimize_CLIP(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ): if show_intermediate: show_pil(lowerCamelCase ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({"Image": wandb.Image(lowerCamelCase )} ) if show_final: show_pil(lowerCamelCase ) if save_final: transformed_img.save(os.path.join(self.save_path , F'''iter_{iter:03d}_final.png''' ) )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ : int = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __magic_name__ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from math import pow def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): '''simple docstring''' if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count _snake_case = int(pow(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n _snake_case , _snake_case = backtrack( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , current_number + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. _snake_case , _snake_case = backtrack( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , current_number + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return current_sum, solutions_count def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not (1 <= needed_sum <= 10_00 and 2 <= power <= 10): raise ValueError( "Invalid input\n" "needed_sum must be between 1 and 1000, power between 2 and 10." ) return backtrack(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' import string def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "" for i in sequence: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if 65 <= extract <= 90: output += chr(1_55 - extract ) elif 97 <= extract <= 1_22: output += chr(2_19 - extract ) else: output += i return output def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = string.ascii_letters _snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(SCREAMING_SNAKE_CASE__ )] if c in letters else c for c in sequence ) def snake_case_ ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) _snake_case = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'''> atbash_slow(): {timeit("atbash_slow(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) print(f'''> atbash(): {timeit("atbash(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'{example} encrypted in atbash: {atbash(example)}') benchmark()
672
1
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def UpperCamelCase( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: _snake_case = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) _snake_case = input_file.read() _snake_case = regexp.search(lowerCamelCase ) return match def UpperCamelCase( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: _snake_case = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) _snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` _snake_case = regexp.finditer(lowerCamelCase ) _snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCamelCase( self ): _snake_case = Path("./datasets" ) _snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F'''open(...) must use utf-8 encoding in {dataset}''' ) def UpperCamelCase( self ): _snake_case = Path("./datasets" ) _snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
672
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' from __future__ import annotations from decimal import Decimal from numpy import array def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(SCREAMING_SNAKE_CASE__ ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix _snake_case = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creates a copy of the matrix with swapped positions of the elements _snake_case = [[0.0, 0.0], [0.0, 0.0]] _snake_case , _snake_case = matrix[1][1], matrix[0][0] _snake_case , _snake_case = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(SCREAMING_SNAKE_CASE__ ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(SCREAMING_SNAKE_CASE__ ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule _snake_case = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError("This matrix has no inverse." ) # Creating cofactor matrix _snake_case = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] _snake_case = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) _snake_case = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) _snake_case = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) _snake_case = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) _snake_case = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) _snake_case = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) _snake_case = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) _snake_case = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) _snake_case = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) _snake_case = array(SCREAMING_SNAKE_CASE__ ) for i in range(3 ): for j in range(3 ): _snake_case = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix _snake_case = array(SCREAMING_SNAKE_CASE__ ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(SCREAMING_SNAKE_CASE__ ) # Calculate the inverse of the matrix return [[float(d(SCREAMING_SNAKE_CASE__ ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError("Please provide a matrix of size 2x2 or 3x3." )
672
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
1
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __magic_name__ : int = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __magic_name__ : Any = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __magic_name__ : Any = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return float((preds == labels).mean() ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__="binary" ): '''simple docstring''' _snake_case = simple_accuracy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = float(fa_score(y_true=SCREAMING_SNAKE_CASE__ , y_pred=SCREAMING_SNAKE_CASE__ , average=SCREAMING_SNAKE_CASE__ ) ) return { "accuracy": acc, "f1": fa, } def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {} for id_pred, label in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = f'''{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}''' _snake_case = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: _snake_case = [(pred, label)] _snake_case , _snake_case = [], [] for question, preds_labels in question_map.items(): _snake_case , _snake_case = zip(*SCREAMING_SNAKE_CASE__ ) _snake_case = fa_score(y_true=SCREAMING_SNAKE_CASE__ , y_pred=SCREAMING_SNAKE_CASE__ , average="macro" ) fas.append(SCREAMING_SNAKE_CASE__ ) _snake_case = int(sum(pred == label for pred, label in preds_labels ) == len(SCREAMING_SNAKE_CASE__ ) ) ems.append(SCREAMING_SNAKE_CASE__ ) _snake_case = float(sum(SCREAMING_SNAKE_CASE__ ) / len(SCREAMING_SNAKE_CASE__ ) ) _snake_case = sum(SCREAMING_SNAKE_CASE__ ) / len(SCREAMING_SNAKE_CASE__ ) _snake_case = float(fa_score(y_true=SCREAMING_SNAKE_CASE__ , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): '''simple docstring''' def UpperCamelCase( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def UpperCamelCase( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCamelCase , lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(lowerCamelCase , lowerCamelCase , fa_avg="macro" ) elif self.config_name == "record": _snake_case = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] _snake_case = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(lowerCamelCase , lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCamelCase , lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCamelCase , lowerCamelCase )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
672
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = DistilBertTokenizer UpperCAmelCase__ : Union[str, Any] = DistilBertTokenizerFast UpperCAmelCase__ : List[str] = True @slow def UpperCamelCase( self ): _snake_case = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _snake_case = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
672
1
'''simple docstring''' import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def snake_case_ ( ): '''simple docstring''' raise RuntimeError("CUDA out of memory." ) class __SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self ): super().__init__() _snake_case = nn.Linear(3 , 4 ) _snake_case = nn.BatchNormad(4 ) _snake_case = nn.Linear(4 , 5 ) def UpperCamelCase( self , lowerCamelCase ): return self.lineara(self.batchnorm(self.lineara(lowerCamelCase ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCamelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase( self ): _snake_case = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase , lowerCamelCase ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _snake_case , _snake_case = mock_training_loop_function("hello" ) self.assertListEqual(lowerCamelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, "hello"] ) def UpperCamelCase( self ): @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(lowerCamelCase ): pass with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("No executable batch size found, reached zero." , cm.exception.args[0] ) def UpperCamelCase( self ): @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("No executable batch size found, reached zero." , cm.exception.args[0] ) def UpperCamelCase( self ): @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function(128 , "hello" , "world" ) self.assertIn("Batch size was passed into `f`" , cm.exception.args[0] ) self.assertIn("`f(arg1='hello', arg2='world')" , cm.exception.args[0] ) def UpperCamelCase( self ): @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase ): raise ValueError("Oops, we had an error!" ) with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("Oops, we had an error!" , cm.exception.args[0] ) @require_cuda def UpperCamelCase( self ): _snake_case = torch.cuda.memory_allocated() _snake_case = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCamelCase ) _snake_case = release_memory(lowerCamelCase ) self.assertEqual(torch.cuda.memory_allocated() , lowerCamelCase )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __magic_name__ : Optional[int] = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[int] = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __magic_name__ : Union[str, Any] = { """configuration_bert""": ["""BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BertConfig""", """BertOnnxConfig"""], """tokenization_bert""": ["""BasicTokenizer""", """BertTokenizer""", """WordpieceTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = ["""BertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : int = [ """BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BertForMaskedLM""", """BertForMultipleChoice""", """BertForNextSentencePrediction""", """BertForPreTraining""", """BertForQuestionAnswering""", """BertForSequenceClassification""", """BertForTokenClassification""", """BertLayer""", """BertLMHeadModel""", """BertModel""", """BertPreTrainedModel""", """load_tf_weights_in_bert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Tuple = [ """TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBertEmbeddings""", """TFBertForMaskedLM""", """TFBertForMultipleChoice""", """TFBertForNextSentencePrediction""", """TFBertForPreTraining""", """TFBertForQuestionAnswering""", """TFBertForSequenceClassification""", """TFBertForTokenClassification""", """TFBertLMHeadModel""", """TFBertMainLayer""", """TFBertModel""", """TFBertPreTrainedModel""", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[Any] = ["""TFBertTokenizer"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : int = [ """FlaxBertForCausalLM""", """FlaxBertForMaskedLM""", """FlaxBertForMultipleChoice""", """FlaxBertForNextSentencePrediction""", """FlaxBertForPreTraining""", """FlaxBertForQuestionAnswering""", """FlaxBertForSequenceClassification""", """FlaxBertForTokenClassification""", """FlaxBertModel""", """FlaxBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : Union[str, Any] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = "backbone." if is_semantic else "" _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', "beit.embeddings.cls_token"), (f'''{prefix}patch_embed.proj.weight''', "beit.embeddings.patch_embeddings.projection.weight"), (f'''{prefix}patch_embed.proj.bias''', "beit.embeddings.patch_embeddings.projection.bias"), (f'''{prefix}pos_embed''', "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): _snake_case = "backbone." if is_semantic else "" # queries, keys and values _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = q_bias _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) _snake_case = gamma_a _snake_case = gamma_a def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = False if "rvlcdip" in checkpoint_url else True _snake_case = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE__ , use_mask_token=SCREAMING_SNAKE_CASE__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _snake_case = 10_24 _snake_case = 40_96 _snake_case = 24 _snake_case = 16 # labels if "rvlcdip" in checkpoint_url: _snake_case = 16 _snake_case = "huggingface/label-files" _snake_case = "rvlcdip-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _snake_case = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="cpu" )["model"] _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) # load HuggingFace model _snake_case = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image _snake_case = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ ) _snake_case = prepare_img() _snake_case = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) _snake_case = outputs.logits # verify logits _snake_case = [1, 16] if "rvlcdip" in checkpoint_url else [1, 1_96, 81_92] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE__ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: if has_lm_head: _snake_case = "dit-base" if "base" in checkpoint_url else "dit-large" else: _snake_case = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) if __name__ == "__main__": __magic_name__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) __magic_name__ : Dict = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
672
1
'''simple docstring''' from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract __magic_name__ : List[str] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = to_pil_image(SCREAMING_SNAKE_CASE__ ) _snake_case , _snake_case = pil_image.size _snake_case = pytesseract.image_to_data(SCREAMING_SNAKE_CASE__ , lang=SCREAMING_SNAKE_CASE__ , output_type="dict" , config=SCREAMING_SNAKE_CASE__ ) _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates _snake_case = [idx for idx, word in enumerate(SCREAMING_SNAKE_CASE__ ) if not word.strip()] _snake_case = [word for idx, word in enumerate(SCREAMING_SNAKE_CASE__ ) if idx not in irrelevant_indices] _snake_case = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE__ ) if idx not in irrelevant_indices] _snake_case = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE__ ) if idx not in irrelevant_indices] _snake_case = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE__ ) if idx not in irrelevant_indices] _snake_case = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format _snake_case = [] for x, y, w, h in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [x, y, x + w, y + h] actual_boxes.append(SCREAMING_SNAKE_CASE__ ) # finally, normalize the bounding boxes _snake_case = [] for box in actual_boxes: normalized_boxes.append(normalize_box(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) assert len(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = "" , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_rescale _snake_case = rescale_value _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD _snake_case = apply_ocr _snake_case = ocr_lang _snake_case = tesseract_config def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) _snake_case = (size["height"], size["width"]) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase=None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = apply_ocr if apply_ocr is not None else self.apply_ocr _snake_case = ocr_lang if ocr_lang is not None else self.ocr_lang _snake_case = tesseract_config if tesseract_config is not None else self.tesseract_config _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("If do_normalize is True, image_mean and image_std must be specified." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , "pytesseract" ) _snake_case = [] _snake_case = [] for image in images: _snake_case , _snake_case = apply_tesseract(lowerCamelCase , lowerCamelCase , lowerCamelCase ) words_batch.append(lowerCamelCase ) boxes_batch.append(lowerCamelCase ) if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = BatchFeature(data={"pixel_values": images} , tensor_type=lowerCamelCase ) if apply_ocr: _snake_case = words_batch _snake_case = boxes_batch return data
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = factor * value _snake_case = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
672
1
'''simple docstring''' from scipy.stats import pearsonr import datasets __magic_name__ : Dict = """ Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ __magic_name__ : int = """ Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ __magic_name__ : Dict = """ @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): '''simple docstring''' def UpperCamelCase( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"] , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): if return_pvalue: _snake_case = pearsonr(lowerCamelCase , lowerCamelCase ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCamelCase , lowerCamelCase )[0] )}
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
672
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __magic_name__ : Optional[int] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"shortest_edge": 256} _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) _snake_case = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
1
'''simple docstring''' import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __magic_name__ : int = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = AlbertTokenizer UpperCAmelCase__ : List[str] = AlbertTokenizerFast UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : int = True def UpperCamelCase( self ): super().setUp() # We have a SentencePiece fixture for testing _snake_case = AlbertTokenizer(lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase( self , lowerCamelCase ): _snake_case = "this is a test" _snake_case = "this is a test" return input_text, output_text def UpperCamelCase( self ): _snake_case = "<pad>" _snake_case = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase ) , lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase ) , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "▁eloquent" ) self.assertEqual(len(lowerCamelCase ) , 30_000 ) def UpperCamelCase( self ): self.assertEqual(self.get_tokenizer().vocab_size , 30_000 ) def UpperCamelCase( self ): if not self.test_rust_tokenizer: return _snake_case = self.get_tokenizer() _snake_case = self.get_rust_tokenizer() _snake_case = "I was born in 92000, and this is falsé." _snake_case = tokenizer.tokenize(lowerCamelCase ) _snake_case = rust_tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) _snake_case = tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) _snake_case = rust_tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) _snake_case = self.get_rust_tokenizer() _snake_case = tokenizer.encode(lowerCamelCase ) _snake_case = rust_tokenizer.encode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = AlbertTokenizer(lowerCamelCase , keep_accents=lowerCamelCase ) _snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCamelCase , ["▁this", "▁is", "▁a", "▁test"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase ) , [48, 25, 21, 1_289] ) _snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCamelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "."] ) _snake_case = tokenizer.convert_tokens_to_ids(lowerCamelCase ) self.assertListEqual(lowerCamelCase , [31, 23, 386, 19, 561, 3_050, 15, 17, 48, 25, 8_256, 18, 1, 9] ) _snake_case = tokenizer.convert_ids_to_tokens(lowerCamelCase ) self.assertListEqual( lowerCamelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "."] , ) def UpperCamelCase( self ): _snake_case = AlbertTokenizer(lowerCamelCase ) _snake_case = tokenizer.encode("sequence builders" ) _snake_case = tokenizer.encode("multi-sequence build" ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def UpperCamelCase( self ): # fmt: off _snake_case = {"attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "input_ids": [[2, 21_970, 13, 5, 6_092, 167, 28, 7_103, 2_153, 673, 8, 7_028, 12_051, 18, 17, 7_103, 2_153, 673, 8, 3_515, 18_684, 8, 4_461, 6, 1_927, 297, 8, 12_060, 2_607, 18, 13, 5, 4_461, 15, 10_538, 38, 8, 135, 15, 822, 58, 15, 993, 10_363, 15, 1_460, 8_005, 4_461, 15, 993, 255, 2_328, 9, 9, 9, 6, 26, 1_112, 816, 3_260, 13, 5, 103, 2_377, 6, 17, 1_112, 816, 2_782, 13, 5, 103, 10_641, 6, 29, 84, 2_512, 2_430, 782, 18_684, 2_761, 19, 808, 2_430, 2_556, 17, 855, 1_480, 9_477, 4_091, 128, 11_712, 15, 7_103, 2_153, 673, 17, 24_883, 9_990, 9, 3], [2, 11_502, 25, 1_006, 20, 782, 8, 11_809, 855, 1_732, 19_393, 18_667, 37, 367, 21_018, 69, 1_854, 34, 11_860, 19_124, 27, 156, 225, 17, 193, 4_141, 19, 65, 9_124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2_231, 886, 2_385, 17_659, 84, 14, 16_792, 1_952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase , model_name="albert-base-v2" , revision="6b6560eaf5ff2e250b00c50f380c5389a9c2d82e" , )
672
'''simple docstring''' import baseaa def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaadecode(SCREAMING_SNAKE_CASE__ ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
672
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
672
1
'''simple docstring''' import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( """The `inpainting.py` script is outdated. Please use directly `from diffusers import""" """ StableDiffusionInpaintPipeline` instead.""" )
672
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
1
'''simple docstring''' import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __magic_name__ : List[str] = argparse.ArgumentParser() parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--txt2img_unclip""", default="""kakaobrain/karlo-v1-alpha""", type=str, required=False, help="""The pretrained txt2img unclip.""", ) __magic_name__ : Optional[Any] = parser.parse_args() __magic_name__ : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __magic_name__ : Optional[Any] = CLIPImageProcessor() __magic_name__ : List[str] = CLIPVisionModelWithProjection.from_pretrained("""openai/clip-vit-large-patch14""") __magic_name__ : str = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
672
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
1
'''simple docstring''' import qiskit def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register _snake_case = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] , [0, 1] ) # Execute the circuit on the qasm simulator _snake_case = qiskit.execute(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , shots=10_00 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : str = single_qubit_measure(2, 2) print(F'Total count for various states are: {counts}')
672
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = text_path elif issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [text_path] _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=("train",) ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for split in splits: _snake_case = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader({"train": text_path} , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader({"train": text_path} , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if split: _snake_case = {split: text_path} else: _snake_case = "train" _snake_case = {"train": text_path, "test": text_path} _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
672
1
'''simple docstring''' __magic_name__ : Optional[Any] = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ __magic_name__ : Optional[Any] = [{"""type""": """code""", """content""": INSTALL_CONTENT}] __magic_name__ : int = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ : Any = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __magic_name__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline __magic_name__ : List[Any] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def UpperCamelCase( self , lowerCamelCase ): if isinstance(lowerCamelCase , lowerCamelCase ): _snake_case = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if len(lowerCamelCase ) == 0 or len(lowerCamelCase ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase ) ) if isinstance(lowerCamelCase , lowerCamelCase ): _snake_case = [sequences] _snake_case = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(__UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' def __init__( self , lowerCamelCase=ZeroShotClassificationArgumentHandler() , *lowerCamelCase , **lowerCamelCase ): _snake_case = args_parser super().__init__(*lowerCamelCase , **lowerCamelCase ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def UpperCamelCase( self ): for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=TruncationStrategy.ONLY_FIRST , **lowerCamelCase ): _snake_case = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _snake_case = self.tokenizer.eos_token try: _snake_case = self.tokenizer( lowerCamelCase , add_special_tokens=lowerCamelCase , return_tensors=lowerCamelCase , padding=lowerCamelCase , truncation=lowerCamelCase , ) except Exception as e: if "too short" in str(lowerCamelCase ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _snake_case = self.tokenizer( lowerCamelCase , add_special_tokens=lowerCamelCase , return_tensors=lowerCamelCase , padding=lowerCamelCase , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def UpperCamelCase( self , **lowerCamelCase ): if kwargs.get("multi_class" , lowerCamelCase ) is not None: _snake_case = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _snake_case = {} if "candidate_labels" in kwargs: _snake_case = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _snake_case = kwargs["hypothesis_template"] _snake_case = {} if "multi_label" in kwargs: _snake_case = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self , lowerCamelCase , *lowerCamelCase , **lowerCamelCase , ): if len(lowerCamelCase ) == 0: pass elif len(lowerCamelCase ) == 1 and "candidate_labels" not in kwargs: _snake_case = args[0] else: raise ValueError(F'''Unable to understand extra arguments {args}''' ) return super().__call__(lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=None , lowerCamelCase="This example is {}." ): _snake_case , _snake_case = self._args_parser(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase , lowerCamelCase ) ): _snake_case = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase ) - 1, **model_input, } def UpperCamelCase( self , lowerCamelCase ): _snake_case = inputs["candidate_label"] _snake_case = inputs["sequence"] _snake_case = {k: inputs[k] for k in self.tokenizer.model_input_names} _snake_case = self.model(**lowerCamelCase ) _snake_case = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=False ): _snake_case = [outputs["candidate_label"] for outputs in model_outputs] _snake_case = [outputs["sequence"] for outputs in model_outputs] _snake_case = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _snake_case = logits.shape[0] _snake_case = len(lowerCamelCase ) _snake_case = N // n _snake_case = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _snake_case = self.entailment_id _snake_case = -1 if entailment_id == 0 else 0 _snake_case = reshaped_outputs[..., [contradiction_id, entailment_id]] _snake_case = np.exp(lowerCamelCase ) / np.exp(lowerCamelCase ).sum(-1 , keepdims=lowerCamelCase ) _snake_case = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _snake_case = reshaped_outputs[..., self.entailment_id] _snake_case = np.exp(lowerCamelCase ) / np.exp(lowerCamelCase ).sum(-1 , keepdims=lowerCamelCase ) _snake_case = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE__ , 2 ) - a def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 2 * x def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2.0 while start <= a: _snake_case = math.pow(SCREAMING_SNAKE_CASE__ , 2 ) return start def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 99_99 , SCREAMING_SNAKE_CASE__ = 0.00000000000001 ): '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _snake_case = get_initial_point(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = value _snake_case = value - fx(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) / fx_derivative(SCREAMING_SNAKE_CASE__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
672
1
'''simple docstring''' import string def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "" for i in sequence: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if 65 <= extract <= 90: output += chr(1_55 - extract ) elif 97 <= extract <= 1_22: output += chr(2_19 - extract ) else: output += i return output def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = string.ascii_letters _snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(SCREAMING_SNAKE_CASE__ )] if c in letters else c for c in sequence ) def snake_case_ ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) _snake_case = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'''> atbash_slow(): {timeit("atbash_slow(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) print(f'''> atbash(): {timeit("atbash(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'{example} encrypted in atbash: {atbash(example)}') benchmark()
672
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[int] = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = '''git_vision_model''' def __init__( self , lowerCamelCase=768 , lowerCamelCase=3_072 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=16 , lowerCamelCase="quick_gelu" , lowerCamelCase=1e-5 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = num_channels _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = '''git''' def __init__( self , lowerCamelCase=None , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=6 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1_024 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=101 , lowerCamelCase=102 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , pad_token_id=lowerCamelCase , **lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) _snake_case = GitVisionConfig(**lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = num_image_with_embedding _snake_case = bos_token_id _snake_case = eos_token_id def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.__class__.model_type return output
672
1
'''simple docstring''' import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Any = '''ssube/stable-diffusion-x4-upscaler-onnx''' def UpperCamelCase( self , lowerCamelCase=0 ): _snake_case = floats_tensor((1, 3, 128, 128) , rng=random.Random(lowerCamelCase ) ) _snake_case = torch.manual_seed(lowerCamelCase ) _snake_case = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCamelCase( self ): _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs() _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) _snake_case = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def UpperCamelCase( self ): _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs() _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array( [0.6898892, 0.59240556, 0.52499527, 0.58866215, 0.52258235, 0.52572715, 0.62414473, 0.6174387, 0.6214964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase( self ): _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs() _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array( [0.7659278, 0.76437664, 0.75579107, 0.7691116, 0.77666986, 0.7727672, 0.7758664, 0.7812226, 0.76942515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase( self ): _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs() _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase( self ): _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs() _snake_case = pipe(**lowerCamelCase ).images _snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array( [0.77424496, 0.773601, 0.7645288, 0.7769598, 0.7772739, 0.7738688, 0.78187233, 0.77879584, 0.767043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase( self ): _snake_case = ort.SessionOptions() _snake_case = False return options def UpperCamelCase( self ): _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) _snake_case = init_image.resize((128, 128) ) # using the PNDM scheduler by default _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = "A fantasy landscape, trending on artstation" _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCamelCase , image=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=10 , generator=lowerCamelCase , output_type="np" , ) _snake_case = output.images _snake_case = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _snake_case = np.array([0.4883, 0.4947, 0.4980, 0.4975, 0.4982, 0.4980, 0.5000, 0.5006, 0.4972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase( self ): _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) _snake_case = init_image.resize((128, 128) ) _snake_case = LMSDiscreteScheduler.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , subfolder="scheduler" ) _snake_case = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , scheduler=lowerCamelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = "A fantasy landscape, trending on artstation" _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCamelCase , image=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=20 , generator=lowerCamelCase , output_type="np" , ) _snake_case = output.images _snake_case = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _snake_case = np.array( [0.50173753, 0.50223356, 0.502039, 0.50233036, 0.5023725, 0.5022601, 0.5018758, 0.50234085, 0.50241566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
672
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __magic_name__ : Dict = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , np.ndarray ): return list(tensor.shape ) _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) if tensor.shape == tf.TensorShape(SCREAMING_SNAKE_CASE__ ): return dynamic _snake_case = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(SCREAMING_SNAKE_CASE__ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise NotImplementedError("Only 1D weight and bias tensors are supported for now, with only a single axis." ) # Get mean and variance on the axis to be normalized _snake_case , _snake_case = tf.nn.moments(SCREAMING_SNAKE_CASE__ , axes=[axis] , keepdims=SCREAMING_SNAKE_CASE__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis _snake_case = [1] * inputs.shape.rank _snake_case = shape_list(SCREAMING_SNAKE_CASE__ )[axis] _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Compute layer normalization using the batch_normalization # function. _snake_case = tf.nn.batch_normalization( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , offset=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , variance_epsilon=SCREAMING_SNAKE_CASE__ , ) return outputs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) _snake_case = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) _snake_case = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ): _snake_case = tf.convert_to_tensor(SCREAMING_SNAKE_CASE__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: _snake_case = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: _snake_case = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) _snake_case = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( SCREAMING_SNAKE_CASE__ , tf.cast(SCREAMING_SNAKE_CASE__ , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(SCREAMING_SNAKE_CASE__ )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. _snake_case = [x for x in data if len(SCREAMING_SNAKE_CASE__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( "The following attributes cannot be saved to HDF5 file because " f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) _snake_case = np.asarray(SCREAMING_SNAKE_CASE__ ) _snake_case = 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(SCREAMING_SNAKE_CASE__ ): _snake_case = chunk_data else: _snake_case = data def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if name in group.attrs: _snake_case = [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs[name]] else: _snake_case = [] _snake_case = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs["%s%d" % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _expand_single_ad_tensor(SCREAMING_SNAKE_CASE__ ): if isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(SCREAMING_SNAKE_CASE__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , SCREAMING_SNAKE_CASE__ )
672
1
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if ( (cp >= 0X4_E_0_0 and cp <= 0X9_F_F_F) or (cp >= 0X3_4_0_0 and cp <= 0X4_D_B_F) # or (cp >= 0X2_0_0_0_0 and cp <= 0X2_A_6_D_F) # or (cp >= 0X2_A_7_0_0 and cp <= 0X2_B_7_3_F) # or (cp >= 0X2_B_7_4_0 and cp <= 0X2_B_8_1_F) # or (cp >= 0X2_B_8_2_0 and cp <= 0X2_C_E_A_F) # or (cp >= 0XF_9_0_0 and cp <= 0XF_A_F_F) or (cp >= 0X2_F_8_0_0 and cp <= 0X2_F_A_1_F) # ): # return True return False def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' for char in word: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if not _is_chinese_char(SCREAMING_SNAKE_CASE__ ): return 0 return 1 def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = set() for token in tokens: _snake_case = len(SCREAMING_SNAKE_CASE__ ) > 1 and is_chinese(SCREAMING_SNAKE_CASE__ ) if chinese_word: word_set.add(SCREAMING_SNAKE_CASE__ ) _snake_case = list(SCREAMING_SNAKE_CASE__ ) return word_list def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not chinese_word_set: return bert_tokens _snake_case = max([len(SCREAMING_SNAKE_CASE__ ) for w in chinese_word_set] ) _snake_case = bert_tokens _snake_case , _snake_case = 0, len(SCREAMING_SNAKE_CASE__ ) while start < end: _snake_case = True if is_chinese(bert_word[start] ): _snake_case = min(end - start , SCREAMING_SNAKE_CASE__ ) for i in range(SCREAMING_SNAKE_CASE__ , 1 , -1 ): _snake_case = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _snake_case = "##" + bert_word[j] _snake_case = start + i _snake_case = False break if single_word: start += 1 return bert_word def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [] for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) , 1_00 ): _snake_case = ltp_tokenizer.seg(lines[i : i + 1_00] )[0] _snake_case = [get_chinese_word(SCREAMING_SNAKE_CASE__ ) for r in res] ltp_res.extend(SCREAMING_SNAKE_CASE__ ) assert len(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ) _snake_case = [] for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) , 1_00 ): _snake_case = bert_tokenizer(lines[i : i + 1_00] , add_special_tokens=SCREAMING_SNAKE_CASE__ , truncation=SCREAMING_SNAKE_CASE__ , max_length=5_12 ) bert_res.extend(res["input_ids"] ) assert len(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ) _snake_case = [] for input_ids, chinese_word in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [] for id in input_ids: _snake_case = bert_tokenizer._convert_id_to_token(SCREAMING_SNAKE_CASE__ ) input_tokens.append(SCREAMING_SNAKE_CASE__ ) _snake_case = add_sub_symbol(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(SCREAMING_SNAKE_CASE__ ): if token[:2] == "##": _snake_case = token[2:] # save chinese tokens' pos if len(SCREAMING_SNAKE_CASE__ ) == 1 and _is_chinese_char(ord(SCREAMING_SNAKE_CASE__ ) ): ref_id.append(SCREAMING_SNAKE_CASE__ ) ref_ids.append(SCREAMING_SNAKE_CASE__ ) assert len(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ) return ref_ids def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' with open(args.file_name , "r" , encoding="utf-8" ) as f: _snake_case = f.readlines() _snake_case = [line.strip() for line in data if len(SCREAMING_SNAKE_CASE__ ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _snake_case = LTP(args.ltp ) # faster in GPU device _snake_case = BertTokenizer.from_pretrained(args.bert ) _snake_case = prepare_ref(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) with open(args.save_path , "w" , encoding="utf-8" ) as f: _snake_case = [json.dumps(SCREAMING_SNAKE_CASE__ ) + "\n" for ref in ref_ids] f.writelines(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : Optional[int] = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __magic_name__ : Optional[Any] = parser.parse_args() main(args)
672
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
1
'''simple docstring''' import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=16 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=3 , lowerCamelCase=4 , lowerCamelCase=None , ): _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_input_mask _snake_case = use_token_type_ids _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = num_labels _snake_case = num_choices _snake_case = scope def UpperCamelCase( self ): _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = None if self.use_input_mask: _snake_case = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case = None if self.use_token_type_ids: _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case = None _snake_case = None _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case = ids_tensor([self.batch_size] , self.num_choices ) _snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase( self ): return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , use_stable_embedding=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = OpenLlamaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase ) _snake_case = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = True _snake_case = OpenLlamaModel(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , ) _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , ) _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = OpenLlamaForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = True _snake_case = True _snake_case = OpenLlamaForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() # first forward pass _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , use_cache=lowerCamelCase , ) _snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) _snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] _snake_case = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] # select random slice _snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() _snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() _snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) def UpperCamelCase( self ): _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = config_and_inputs _snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) UpperCAmelCase__ : Dict = (OpenLlamaForCausalLM,) if is_torch_available() else () UpperCAmelCase__ : Dict = ( { '''feature-extraction''': OpenLlamaModel, '''text-classification''': OpenLlamaForSequenceClassification, '''text-generation''': OpenLlamaForCausalLM, '''zero-shot''': OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : int = False UpperCAmelCase__ : Dict = False def UpperCamelCase( self ): _snake_case = OpenLlamaModelTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase , hidden_size=37 ) def UpperCamelCase( self ): self.config_tester.run_common_tests() def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _snake_case = type self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = 3 _snake_case = input_dict["input_ids"] _snake_case = input_ids.ne(1 ).to(lowerCamelCase ) _snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _snake_case = OpenLlamaForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = 3 _snake_case = "single_label_classification" _snake_case = input_dict["input_ids"] _snake_case = input_ids.ne(1 ).to(lowerCamelCase ) _snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _snake_case = OpenLlamaForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = 3 _snake_case = "multi_label_classification" _snake_case = input_dict["input_ids"] _snake_case = input_ids.ne(1 ).to(lowerCamelCase ) _snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _snake_case = OpenLlamaForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() _snake_case = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("Open-Llama buffers include complex numbers, which breaks this test" ) def UpperCamelCase( self ): pass @parameterized.expand([("linear",), ("dynamic",)] ) def UpperCamelCase( self , lowerCamelCase ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = ids_tensor([1, 10] , config.vocab_size ) _snake_case = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _snake_case = OpenLlamaModel(lowerCamelCase ) original_model.to(lowerCamelCase ) original_model.eval() _snake_case = original_model(lowerCamelCase ).last_hidden_state _snake_case = original_model(lowerCamelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _snake_case = {"type": scaling_type, "factor": 10.0} _snake_case = OpenLlamaModel(lowerCamelCase ) scaled_model.to(lowerCamelCase ) scaled_model.eval() _snake_case = scaled_model(lowerCamelCase ).last_hidden_state _snake_case = scaled_model(lowerCamelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-5 ) )
672
'''simple docstring''' from torch import nn def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'''Unsupported activation function: {act_fn}''' )
672
1
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return credit_card_number.startswith(("34", "35", "37", "4", "5", "6") ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = credit_card_number _snake_case = 0 _snake_case = len(SCREAMING_SNAKE_CASE__ ) - 2 for i in range(SCREAMING_SNAKE_CASE__ , -1 , -2 ): # double the value of every second digit _snake_case = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 _snake_case = cc_number[:i] + str(SCREAMING_SNAKE_CASE__ ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(SCREAMING_SNAKE_CASE__ ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = f'''{credit_card_number} is an invalid credit card number because''' if not credit_card_number.isdigit(): print(f'''{error_message} it has nonnumerical characters.''' ) return False if not 13 <= len(SCREAMING_SNAKE_CASE__ ) <= 16: print(f'''{error_message} of its length.''' ) return False if not validate_initial_digits(SCREAMING_SNAKE_CASE__ ): print(f'''{error_message} of its first two digits.''' ) return False if not luhn_validation(SCREAMING_SNAKE_CASE__ ): print(f'''{error_message} it fails the Luhn check.''' ) return False print(f'''{credit_card_number} is a valid credit card number.''' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("""4111111111111111""") validate_credit_card_number("""32323""")
672
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __magic_name__ : Tuple = 0 __magic_name__ : Dict = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __magic_name__ : str = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __magic_name__ : Dict = tuple[int, int] class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): _snake_case = pos_x _snake_case = pos_y _snake_case = (pos_y, pos_x) _snake_case = goal_x _snake_case = goal_y _snake_case = g_cost _snake_case = parent _snake_case = self.calculate_heuristic() _snake_case = self.g_cost + self.h_cost def UpperCamelCase( self ): _snake_case = self.pos_x - self.goal_x _snake_case = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCamelCase ) + abs(lowerCamelCase ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowerCamelCase ): return self.f_cost < other.f_cost class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCamelCase ) _snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , lowerCamelCase ) _snake_case = [self.start] _snake_case = [] _snake_case = False def UpperCamelCase( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _snake_case = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCamelCase ) self.closed_nodes.append(lowerCamelCase ) _snake_case = self.get_successors(lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = self.open_nodes.pop(self.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCamelCase ) else: self.open_nodes.append(lowerCamelCase ) return [self.start.pos] def UpperCamelCase( self , lowerCamelCase ): _snake_case = [] for action in delta: _snake_case = parent.pos_x + action[1] _snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCamelCase , lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCamelCase , ) ) return successors def UpperCamelCase( self , lowerCamelCase ): _snake_case = node _snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _snake_case = current_node.parent path.reverse() return path class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase ): _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = AStar(lowerCamelCase , lowerCamelCase ) _snake_case = False def UpperCamelCase( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _snake_case = self.fwd_astar.open_nodes.pop(0 ) _snake_case = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCamelCase , lowerCamelCase ) self.fwd_astar.closed_nodes.append(lowerCamelCase ) self.bwd_astar.closed_nodes.append(lowerCamelCase ) _snake_case = current_bwd_node _snake_case = current_fwd_node _snake_case = { self.fwd_astar: self.fwd_astar.get_successors(lowerCamelCase ), self.bwd_astar: self.bwd_astar.get_successors(lowerCamelCase ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCamelCase ) else: # retrieve the best current path _snake_case = astar.open_nodes.pop( astar.open_nodes.index(lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCamelCase ) else: astar.open_nodes.append(lowerCamelCase ) return [self.fwd_astar.start.pos] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): _snake_case = self.fwd_astar.retrace_path(lowerCamelCase ) _snake_case = self.bwd_astar.retrace_path(lowerCamelCase ) bwd_path.pop() bwd_path.reverse() _snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __magic_name__ : Optional[int] = (0, 0) __magic_name__ : Any = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __magic_name__ : Any = time.time() __magic_name__ : Optional[int] = AStar(init, goal) __magic_name__ : str = a_star.search() __magic_name__ : List[Any] = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') __magic_name__ : List[str] = time.time() __magic_name__ : Optional[Any] = BidirectionalAStar(init, goal) __magic_name__ : Optional[int] = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
672
1
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __magic_name__ : str = random.Random() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1.0 , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None ): '''simple docstring''' if rng is None: _snake_case = global_rng _snake_case = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=400 , lowerCamelCase=2_000 , lowerCamelCase=1 , lowerCamelCase=0.0 , lowerCamelCase=16_000 , lowerCamelCase=True , lowerCamelCase=True , ): _snake_case = parent _snake_case = batch_size _snake_case = min_seq_length _snake_case = max_seq_length _snake_case = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _snake_case = feature_size _snake_case = padding_value _snake_case = sampling_rate _snake_case = return_attention_mask _snake_case = do_normalize def UpperCamelCase( self ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase( self , lowerCamelCase=False , lowerCamelCase=False ): def _flatten(lowerCamelCase ): return list(itertools.chain(*lowerCamelCase ) ) if equal_length: _snake_case = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size _snake_case = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _snake_case = [np.asarray(lowerCamelCase ) for x in speech_inputs] return speech_inputs class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = WavaVecaFeatureExtractor def UpperCamelCase( self ): _snake_case = WavaVecaFeatureExtractionTester(self ) def UpperCamelCase( self , lowerCamelCase ): self.assertTrue(np.all(np.mean(lowerCamelCase , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _snake_case = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case = [np.asarray(lowerCamelCase ) for speech_input in speech_inputs] # Test not batched input _snake_case = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values _snake_case = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) # Test batched _snake_case = feat_extract(lowerCamelCase , return_tensors="np" ).input_values _snake_case = feat_extract(lowerCamelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase , lowerCamelCase ): self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _snake_case = [floats_list((1, x) )[0] for x in (800, 800, 800)] _snake_case = np.asarray(lowerCamelCase ) _snake_case = feat_extract(lowerCamelCase , return_tensors="np" ).input_values _snake_case = feat_extract(lowerCamelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCamelCase , lowerCamelCase ): self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) def UpperCamelCase( self ): _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _snake_case = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case = ["longest", "max_length", "do_not_pad"] _snake_case = [None, 1_600, None] for max_length, padding in zip(lowerCamelCase , lowerCamelCase ): _snake_case = feat_extract(lowerCamelCase , padding=lowerCamelCase , max_length=lowerCamelCase , return_tensors="np" ) _snake_case = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self.assertTrue(input_values[0][1_000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def UpperCamelCase( self ): _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _snake_case = range(800 , 1_400 , 200 ) _snake_case = [floats_list((1, x) )[0] for x in lengths] _snake_case = ["longest", "max_length", "do_not_pad"] _snake_case = [None, 1_600, None] for max_length, padding in zip(lowerCamelCase , lowerCamelCase ): _snake_case = feat_extract(lowerCamelCase , max_length=lowerCamelCase , padding=lowerCamelCase ) _snake_case = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def UpperCamelCase( self ): _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _snake_case = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=1_000 , padding="max_length" , return_tensors="np" ) _snake_case = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCamelCase( self ): _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _snake_case = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=1_000 , padding="longest" , return_tensors="np" ) _snake_case = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1_000) ) _snake_case = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case = feat_extract( lowerCamelCase , truncation=lowerCamelCase , max_length=2_000 , padding="longest" , return_tensors="np" ) _snake_case = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1_200) ) @require_torch def UpperCamelCase( self ): import torch _snake_case = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _snake_case = np.random.rand(100 ).astype(np.floataa ) _snake_case = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _snake_case = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) _snake_case = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCamelCase( self ): # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: _snake_case = WavaVecaConfig.from_pretrained(lowerCamelCase ) _snake_case = WavaVecaFeatureExtractor.from_pretrained(lowerCamelCase ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == "layer" )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ : int = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __magic_name__ : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import logging from transformers.configuration_utils import PretrainedConfig __magic_name__ : Any = logging.getLogger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = '''masked_bert''' def __init__( self , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="topK" , lowerCamelCase="constant" , lowerCamelCase=0.0 , **lowerCamelCase , ): super().__init__(pad_token_id=lowerCamelCase , **lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = pruning_method _snake_case = mask_init _snake_case = mask_scale
672
'''simple docstring''' import string def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = "" for i in sequence: _snake_case = ord(SCREAMING_SNAKE_CASE__ ) if 65 <= extract <= 90: output += chr(1_55 - extract ) elif 97 <= extract <= 1_22: output += chr(2_19 - extract ) else: output += i return output def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = string.ascii_letters _snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(SCREAMING_SNAKE_CASE__ )] if c in letters else c for c in sequence ) def snake_case_ ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) _snake_case = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'''> atbash_slow(): {timeit("atbash_slow(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) print(f'''> atbash(): {timeit("atbash(printable)" , setup=SCREAMING_SNAKE_CASE__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F'{example} encrypted in atbash: {atbash(example)}') benchmark()
672
1
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
'''simple docstring''' import numpy as np def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 1 / (1 + np.exp(-vector )) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( """The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion""" ) __magic_name__ : List[str] = None __magic_name__ : str = { """7B""": 11_008, """13B""": 13_824, """30B""": 17_920, """65B""": 22_016, """70B""": 28_672, } __magic_name__ : Union[str, Any] = { """7B""": 1, """7Bf""": 1, """13B""": 2, """13Bf""": 2, """30B""": 4, """65B""": 8, """70B""": 8, """70Bf""": 8, } def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , SCREAMING_SNAKE_CASE__=2_56 ): '''simple docstring''' return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE__ , "r" ) as f: return json.load(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE__ , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=True ): '''simple docstring''' os.makedirs(SCREAMING_SNAKE_CASE__ , exist_ok=SCREAMING_SNAKE_CASE__ ) _snake_case = os.path.join(SCREAMING_SNAKE_CASE__ , "tmp" ) os.makedirs(SCREAMING_SNAKE_CASE__ , exist_ok=SCREAMING_SNAKE_CASE__ ) _snake_case = read_json(os.path.join(SCREAMING_SNAKE_CASE__ , "params.json" ) ) _snake_case = NUM_SHARDS[model_size] _snake_case = params["n_layers"] _snake_case = params["n_heads"] _snake_case = n_heads // num_shards _snake_case = params["dim"] _snake_case = dim // n_heads _snake_case = 10000.0 _snake_case = 1.0 / (base ** (torch.arange(0 , SCREAMING_SNAKE_CASE__ , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: _snake_case = params["n_kv_heads"] # for GQA / MQA _snake_case = n_heads_per_shard // num_key_value_heads _snake_case = dim // num_key_value_heads else: # compatibility with other checkpoints _snake_case = n_heads _snake_case = n_heads_per_shard _snake_case = dim # permute for sliced rotary def permute(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=n_heads , SCREAMING_SNAKE_CASE__=dim , SCREAMING_SNAKE_CASE__=dim ): return w.view(SCREAMING_SNAKE_CASE__ , dima // n_heads // 2 , 2 , SCREAMING_SNAKE_CASE__ ).transpose(1 , 2 ).reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) print(f'''Fetching all parameters from the checkpoint at {input_base_path}.''' ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) _snake_case = torch.load(os.path.join(SCREAMING_SNAKE_CASE__ , "consolidated.00.pth" ) , map_location="cpu" ) else: # Sharded _snake_case = [ torch.load(os.path.join(SCREAMING_SNAKE_CASE__ , f'''consolidated.{i:02d}.pth''' ) , map_location="cpu" ) for i in range(SCREAMING_SNAKE_CASE__ ) ] _snake_case = 0 _snake_case = {"weight_map": {}} for layer_i in range(SCREAMING_SNAKE_CASE__ ): _snake_case = f'''pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin''' if model_size == "7B": # Unsharded _snake_case = { f'''model.layers.{layer_i}.self_attn.q_proj.weight''': permute( loaded[f'''layers.{layer_i}.attention.wq.weight'''] ), f'''model.layers.{layer_i}.self_attn.k_proj.weight''': permute( loaded[f'''layers.{layer_i}.attention.wk.weight'''] ), f'''model.layers.{layer_i}.self_attn.v_proj.weight''': loaded[f'''layers.{layer_i}.attention.wv.weight'''], f'''model.layers.{layer_i}.self_attn.o_proj.weight''': loaded[f'''layers.{layer_i}.attention.wo.weight'''], f'''model.layers.{layer_i}.mlp.gate_proj.weight''': loaded[f'''layers.{layer_i}.feed_forward.w1.weight'''], f'''model.layers.{layer_i}.mlp.down_proj.weight''': loaded[f'''layers.{layer_i}.feed_forward.w2.weight'''], f'''model.layers.{layer_i}.mlp.up_proj.weight''': loaded[f'''layers.{layer_i}.feed_forward.w3.weight'''], f'''model.layers.{layer_i}.input_layernorm.weight''': loaded[f'''layers.{layer_i}.attention_norm.weight'''], f'''model.layers.{layer_i}.post_attention_layernorm.weight''': loaded[f'''layers.{layer_i}.ffn_norm.weight'''], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. _snake_case = { f'''model.layers.{layer_i}.input_layernorm.weight''': loaded[0][ f'''layers.{layer_i}.attention_norm.weight''' ].clone(), f'''model.layers.{layer_i}.post_attention_layernorm.weight''': loaded[0][ f'''layers.{layer_i}.ffn_norm.weight''' ].clone(), } _snake_case = permute( torch.cat( [ loaded[i][f'''layers.{layer_i}.attention.wq.weight'''].view(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for i in range(SCREAMING_SNAKE_CASE__ ) ] , dim=0 , ).reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) _snake_case = permute( torch.cat( [ loaded[i][f'''layers.{layer_i}.attention.wk.weight'''].view( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for i in range(SCREAMING_SNAKE_CASE__ ) ] , dim=0 , ).reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) _snake_case = torch.cat( [ loaded[i][f'''layers.{layer_i}.attention.wv.weight'''].view( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for i in range(SCREAMING_SNAKE_CASE__ ) ] , dim=0 , ).reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = torch.cat( [loaded[i][f'''layers.{layer_i}.attention.wo.weight'''] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=1 ) _snake_case = torch.cat( [loaded[i][f'''layers.{layer_i}.feed_forward.w1.weight'''] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=0 ) _snake_case = torch.cat( [loaded[i][f'''layers.{layer_i}.feed_forward.w2.weight'''] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=1 ) _snake_case = torch.cat( [loaded[i][f'''layers.{layer_i}.feed_forward.w3.weight'''] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=0 ) _snake_case = inv_freq for k, v in state_dict.items(): _snake_case = filename param_count += v.numel() torch.save(SCREAMING_SNAKE_CASE__ , os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) _snake_case = f'''pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin''' if model_size == "7B": # Unsharded _snake_case = { "model.embed_tokens.weight": loaded["tok_embeddings.weight"], "model.norm.weight": loaded["norm.weight"], "lm_head.weight": loaded["output.weight"], } else: _snake_case = { "model.norm.weight": loaded[0]["norm.weight"], "model.embed_tokens.weight": torch.cat( [loaded[i]["tok_embeddings.weight"] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=1 ), "lm_head.weight": torch.cat([loaded[i]["output.weight"] for i in range(SCREAMING_SNAKE_CASE__ )] , dim=0 ), } for k, v in state_dict.items(): _snake_case = filename param_count += v.numel() torch.save(SCREAMING_SNAKE_CASE__ , os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) # Write configs _snake_case = {"total_size": param_count * 2} write_json(SCREAMING_SNAKE_CASE__ , os.path.join(SCREAMING_SNAKE_CASE__ , "pytorch_model.bin.index.json" ) ) _snake_case = params["ffn_dim_multiplier"] if "ffn_dim_multiplier" in params else 1 _snake_case = params["multiple_of"] if "multiple_of" in params else 2_56 _snake_case = LlamaConfig( hidden_size=SCREAMING_SNAKE_CASE__ , intermediate_size=compute_intermediate_size(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , num_attention_heads=params["n_heads"] , num_hidden_layers=params["n_layers"] , rms_norm_eps=params["norm_eps"] , num_key_value_heads=SCREAMING_SNAKE_CASE__ , ) config.save_pretrained(SCREAMING_SNAKE_CASE__ ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print("Loading the checkpoint in a Llama model." ) _snake_case = LlamaForCausalLM.from_pretrained(SCREAMING_SNAKE_CASE__ , torch_dtype=torch.floataa , low_cpu_mem_usage=SCREAMING_SNAKE_CASE__ ) # Avoid saving this as part of the config. del model.config._name_or_path print("Saving in the Transformers format." ) model.save_pretrained(SCREAMING_SNAKE_CASE__ , safe_serialization=SCREAMING_SNAKE_CASE__ ) shutil.rmtree(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f'''Saving a {tokenizer_class.__name__} to {tokenizer_path}.''' ) _snake_case = tokenizer_class(SCREAMING_SNAKE_CASE__ ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( ): '''simple docstring''' _snake_case = argparse.ArgumentParser() parser.add_argument( "--input_dir" , help="Location of LLaMA weights, which contains tokenizer.model and model folders" , ) parser.add_argument( "--model_size" , choices=["7B", "7Bf", "13B", "13Bf", "30B", "65B", "70B", "70Bf", "tokenizer_only"] , ) parser.add_argument( "--output_dir" , help="Location to write HF model and tokenizer" , ) parser.add_argument("--safe_serialization" , type=SCREAMING_SNAKE_CASE__ , help="Whether or not to save using `safetensors`." ) _snake_case = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) _snake_case = os.path.join(args.input_dir , "tokenizer.model" ) write_tokenizer(args.output_dir , SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": main()
672
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
1
'''simple docstring''' from __future__ import annotations from collections.abc import Sequence from typing import Literal def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = list(SCREAMING_SNAKE_CASE__ ) _snake_case = list(SCREAMING_SNAKE_CASE__ ) _snake_case = 0 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): if lista[i] != lista[i]: count += 1 _snake_case = "_" if count > 1: return False else: return "".join(SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [] while True: _snake_case = ["$"] * len(SCREAMING_SNAKE_CASE__ ) _snake_case = [] for i in range(len(SCREAMING_SNAKE_CASE__ ) ): for j in range(i + 1 , len(SCREAMING_SNAKE_CASE__ ) ): _snake_case = compare_string(binary[i] , binary[j] ) if k is False: _snake_case = "*" _snake_case = "*" temp.append("X" ) for i in range(len(SCREAMING_SNAKE_CASE__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(SCREAMING_SNAKE_CASE__ ) == 0: return pi _snake_case = list(set(SCREAMING_SNAKE_CASE__ ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [] for minterm in minterms: _snake_case = "" for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = str(minterm % 2 ) + string minterm //= 2 temp.append(SCREAMING_SNAKE_CASE__ ) return temp def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = list(SCREAMING_SNAKE_CASE__ ) _snake_case = list(SCREAMING_SNAKE_CASE__ ) _snake_case = 0 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [] _snake_case = [0] * len(SCREAMING_SNAKE_CASE__ ) for i in range(len(chart[0] ) ): _snake_case = 0 _snake_case = -1 for j in range(len(SCREAMING_SNAKE_CASE__ ) ): if chart[j][i] == 1: count += 1 _snake_case = j if count == 1: _snake_case = 1 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(SCREAMING_SNAKE_CASE__ ) ): _snake_case = 0 temp.append(prime_implicants[i] ) while True: _snake_case = 0 _snake_case = -1 _snake_case = 0 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): _snake_case = chart[i].count(1 ) if count_n > max_n: _snake_case = count_n _snake_case = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(SCREAMING_SNAKE_CASE__ ) ): _snake_case = 0 def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = [[0 for x in range(len(SCREAMING_SNAKE_CASE__ ) )] for x in range(len(SCREAMING_SNAKE_CASE__ ) )] for i in range(len(SCREAMING_SNAKE_CASE__ ) ): _snake_case = prime_implicants[i].count("_" ) for j in range(len(SCREAMING_SNAKE_CASE__ ) ): if is_for_table(prime_implicants[i] , binary[j] , SCREAMING_SNAKE_CASE__ ): _snake_case = 1 return chart def snake_case_ ( ): '''simple docstring''' _snake_case = int(input("Enter the no. of variables\n" ) ) _snake_case = [ float(SCREAMING_SNAKE_CASE__ ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] _snake_case = decimal_to_binary(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = check(SCREAMING_SNAKE_CASE__ ) print("Prime Implicants are:" ) print(SCREAMING_SNAKE_CASE__ ) _snake_case = prime_implicant_chart(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = selection(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) print("Essential Prime Implicants are:" ) print(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
672
'''simple docstring''' from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = DistilBertTokenizer UpperCAmelCase__ : Union[str, Any] = DistilBertTokenizerFast UpperCAmelCase__ : List[str] = True @slow def UpperCamelCase( self ): _snake_case = DistilBertTokenizer.from_pretrained("distilbert-base-uncased" ) _snake_case = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) _snake_case = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
672
1
'''simple docstring''' import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = 10 def UpperCamelCase( self ): _snake_case = [1, 2, 3, 4] _snake_case = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(lowerCamelCase , self.block_size , 0 ) , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] _snake_case = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(lowerCamelCase , self.block_size , 0 ) , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] _snake_case = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(lowerCamelCase , self.block_size , 0 ) , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = "It was the year of Our Lord one thousand seven hundred and\n seventy-five.\n\nSpiritual revelations were conceded to England at that\n favoured period, as at this." _snake_case , _snake_case = process_story(lowerCamelCase ) self.assertEqual(lowerCamelCase , [] ) def UpperCamelCase( self ): _snake_case = "" _snake_case , _snake_case = process_story(lowerCamelCase ) self.assertEqual(lowerCamelCase , [] ) self.assertEqual(lowerCamelCase , [] ) def UpperCamelCase( self ): _snake_case = ( "It was the year of Our Lord one thousand seven hundred and " "seventy-five\n\nSpiritual revelations were conceded to England " "at that favoured period, as at this.\n@highlight\n\nIt was the best of times" ) _snake_case , _snake_case = process_story(lowerCamelCase ) _snake_case = [ "It was the year of Our Lord one thousand seven hundred and seventy-five.", "Spiritual revelations were conceded to England at that favoured period, as at this.", ] self.assertEqual(lowerCamelCase , lowerCamelCase ) _snake_case = ["It was the best of times."] self.assertEqual(lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = torch.tensor([1, 2, 3, 4] ) _snake_case = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(lowerCamelCase , 0 ).numpy() , expected.numpy() ) def UpperCamelCase( self ): _snake_case = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) _snake_case = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowerCamelCase , 23 ).numpy() , expected.numpy() ) def UpperCamelCase( self ): _snake_case = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) _snake_case = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowerCamelCase , 1 ).numpy() , expected.numpy() ) def UpperCamelCase( self ): _snake_case = 101 _snake_case = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]] ) _snake_case = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) _snake_case = compute_token_type_ids(lowerCamelCase , lowerCamelCase ) np.testing.assert_array_equal(lowerCamelCase , lowerCamelCase )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __magic_name__ : Optional[int] = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[int] = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __magic_name__ : str = """platform""" import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , ): '''simple docstring''' if attention_mask is None: _snake_case = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=99 , lowerCamelCase=16 , lowerCamelCase=2 , lowerCamelCase=4 , lowerCamelCase=4 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=32 , lowerCamelCase=2 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0.02 , ): _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = initializer_range def UpperCamelCase( self ): _snake_case = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _snake_case = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _snake_case = shift_tokens_right(lowerCamelCase , 1 , 2 ) _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCamelCase , ) _snake_case = prepare_blenderbot_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, inputs_dict def UpperCamelCase( self ): _snake_case , _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = 20 _snake_case = model_class_name(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] ) _snake_case , _snake_case = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase , lowerCamelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCamelCase , ) _snake_case = model.decode(lowerCamelCase , lowerCamelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = 20 _snake_case = model_class_name(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] ) _snake_case , _snake_case = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase , lowerCamelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = model.decode(lowerCamelCase , lowerCamelCase , decoder_attention_mask=lowerCamelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) @require_flax class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Any = 99 def UpperCamelCase( self ): _snake_case = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _snake_case = input_ids.shape[0] _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def UpperCamelCase( self ): _snake_case , _snake_case , _snake_case = self._get_config_and_data() _snake_case = FlaxBlenderbotForConditionalGeneration(lowerCamelCase ) _snake_case = lm_model(input_ids=lowerCamelCase ) _snake_case = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _snake_case = FlaxBlenderbotForConditionalGeneration(lowerCamelCase ) _snake_case = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _snake_case = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _snake_case = lm_model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) _snake_case = (*summary.shape, config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCamelCase ) def UpperCamelCase( self ): _snake_case = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _snake_case = shift_tokens_right(lowerCamelCase , 1 , 2 ) _snake_case = np.equal(lowerCamelCase , 1 ).astype(np.floataa ).sum() _snake_case = np.equal(lowerCamelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCamelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : Any = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) UpperCAmelCase__ : Any = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def UpperCamelCase( self ): _snake_case = FlaxBlenderbotModelTester(self ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) _snake_case = model_class(lowerCamelCase ) @jax.jit def encode_jitted(lowerCamelCase , lowerCamelCase=None , **lowerCamelCase ): return model.encode(input_ids=lowerCamelCase , attention_mask=lowerCamelCase ) with self.subTest("JIT Enabled" ): _snake_case = encode_jitted(**lowerCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _snake_case = encode_jitted(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for jitted_output, output in zip(lowerCamelCase , lowerCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) _snake_case = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowerCamelCase , lowerCamelCase , lowerCamelCase ): return model.decode( decoder_input_ids=lowerCamelCase , decoder_attention_mask=lowerCamelCase , encoder_outputs=lowerCamelCase , ) with self.subTest("JIT Enabled" ): _snake_case = decode_jitted(**lowerCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _snake_case = decode_jitted(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for jitted_output, output in zip(lowerCamelCase , lowerCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCamelCase( self ): for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("facebook/blenderbot-400M-distill" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case = np.ones((1, 1) ) * model.config.eos_token_id _snake_case = model(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @unittest.skipUnless(jax_device != "cpu" , "3B test too slow on CPU." ) @slow def UpperCamelCase( self ): _snake_case = {"num_beams": 1, "early_stopping": True, "min_length": 15, "max_length": 25} _snake_case = {"skip_special_tokens": True, "clean_up_tokenization_spaces": True} _snake_case = FlaxBlenderbotForConditionalGeneration.from_pretrained("facebook/blenderbot-3B" , from_pt=lowerCamelCase ) _snake_case = BlenderbotTokenizer.from_pretrained("facebook/blenderbot-3B" ) _snake_case = ["Sam"] _snake_case = tokenizer(lowerCamelCase , return_tensors="jax" ) _snake_case = model.generate(**lowerCamelCase , **lowerCamelCase ) _snake_case = "Sam is a great name. It means \"sun\" in Gaelic." _snake_case = tokenizer.batch_decode(lowerCamelCase , **lowerCamelCase ) assert generated_txt[0].strip() == tgt_text
672
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : Union[str, Any] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = "backbone." if is_semantic else "" _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''{prefix}blocks.{i}.norm1.weight''', f'''beit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm1.bias''', f'''beit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.weight''', f'''beit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''{prefix}blocks.{i}.attn.proj.bias''', f'''beit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.weight''', f'''beit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.norm2.bias''', f'''beit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.weight''', f'''beit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc1.bias''', f'''beit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.weight''', f'''beit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''{prefix}blocks.{i}.mlp.fc2.bias''', f'''beit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ (f'''{prefix}cls_token''', "beit.embeddings.cls_token"), (f'''{prefix}patch_embed.proj.weight''', "beit.embeddings.patch_embeddings.projection.weight"), (f'''{prefix}patch_embed.proj.bias''', "beit.embeddings.patch_embeddings.projection.bias"), (f'''{prefix}pos_embed''', "beit.embeddings.position_embeddings"), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): _snake_case = "backbone." if is_semantic else "" # queries, keys and values _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.q_bias''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.attn.v_bias''' ) _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = q_bias _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_1''' ) _snake_case = state_dict.pop(f'''{prefix}blocks.{i}.gamma_2''' ) _snake_case = gamma_a _snake_case = gamma_a def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = False if "rvlcdip" in checkpoint_url else True _snake_case = BeitConfig(use_absolute_position_embeddings=SCREAMING_SNAKE_CASE__ , use_mask_token=SCREAMING_SNAKE_CASE__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _snake_case = 10_24 _snake_case = 40_96 _snake_case = 24 _snake_case = 16 # labels if "rvlcdip" in checkpoint_url: _snake_case = 16 _snake_case = "huggingface/label-files" _snake_case = "rvlcdip-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _snake_case = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="cpu" )["model"] _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , has_lm_head=SCREAMING_SNAKE_CASE__ ) # load HuggingFace model _snake_case = BeitForMaskedImageModeling(SCREAMING_SNAKE_CASE__ ) if has_lm_head else BeitForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image _snake_case = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ ) _snake_case = prepare_img() _snake_case = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) _snake_case = outputs.logits # verify logits _snake_case = [1, 16] if "rvlcdip" in checkpoint_url else [1, 1_96, 81_92] assert logits.shape == torch.Size(SCREAMING_SNAKE_CASE__ ), "Shape of logits not as expected" Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: if has_lm_head: _snake_case = "dit-base" if "base" in checkpoint_url else "dit-large" else: _snake_case = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) if __name__ == "__main__": __magic_name__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) __magic_name__ : Dict = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
672
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase( self ): _snake_case = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) _snake_case = { "input_ids": tf.convert_to_tensor([[0, 2_646, 10_269, 83, 99_942, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } _snake_case = model(lowerCamelCase )["last_hidden_state"] _snake_case = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , lowerCamelCase ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [ [ [0.0681762, 0.10894451, 0.06772504], [-0.06423668, 0.02366615, 0.04329344], [-0.06057295, 0.09974135, -0.00070584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False _snake_case = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1 , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = factor * value _snake_case = value while not is_prime(SCREAMING_SNAKE_CASE__ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE__ ) return value
672
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer __magic_name__ : Dict = logging.get_logger(__name__) __magic_name__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} __magic_name__ : int = { """vocab_file""": { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt""" ), } } __magic_name__ : List[Any] = { """junnyu/roformer_chinese_small""": 1_536, """junnyu/roformer_chinese_base""": 1_536, """junnyu/roformer_chinese_char_small""": 512, """junnyu/roformer_chinese_char_base""": 512, """junnyu/roformer_small_discriminator""": 128, """junnyu/roformer_small_generator""": 128, } __magic_name__ : Optional[int] = { """junnyu/roformer_chinese_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_base""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_base""": {"""do_lower_case""": True}, """junnyu/roformer_small_discriminator""": {"""do_lower_case""": True}, """junnyu/roformer_small_generator""": {"""do_lower_case""": True}, } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = VOCAB_FILES_NAMES UpperCAmelCase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : List[str] = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ : Dict = RoFormerTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase="[UNK]" , lowerCamelCase="[SEP]" , lowerCamelCase="[PAD]" , lowerCamelCase="[CLS]" , lowerCamelCase="[MASK]" , lowerCamelCase=True , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , do_lower_case=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , pad_token=lowerCamelCase , cls_token=lowerCamelCase , mask_token=lowerCamelCase , tokenize_chinese_chars=lowerCamelCase , strip_accents=lowerCamelCase , **lowerCamelCase , ) _snake_case = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get("lowercase" , lowerCamelCase ) != do_lower_case or pre_tok_state.get("strip_accents" , lowerCamelCase ) != strip_accents ): _snake_case = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) _snake_case = do_lower_case _snake_case = strip_accents _snake_case = pre_tok_class(**lowerCamelCase ) _snake_case = do_lower_case def __getstate__( self ): _snake_case = self.__dict__.copy() _snake_case = BertPreTokenizer() return state def __setstate__( self , lowerCamelCase ): _snake_case = d _snake_case = self.__dict__["_tokenizer"].get_vocab() _snake_case = PreTokenizer.custom(JiebaPreTokenizer(lowerCamelCase ) ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=None ): _snake_case = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = [self.sep_token_id] _snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=False , **lowerCamelCase , ): _snake_case = BertPreTokenizer() return super().save_pretrained(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase )
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Dict = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[Any] = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys __magic_name__ : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : int = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" _snake_case = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _snake_case = "" else: _snake_case = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = in_proj_bias[: config.hidden_size] _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = in_proj_bias[-config.hidden_size :] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = DeiTConfig() # all deit models have fine-tuned heads _snake_case = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size _snake_case = 10_00 _snake_case = "huggingface/label-files" _snake_case = "imagenet-1k-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} _snake_case = int(deit_name[-6:-4] ) _snake_case = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): _snake_case = 1_92 _snake_case = 7_68 _snake_case = 12 _snake_case = 3 elif deit_name[9:].startswith("small" ): _snake_case = 3_84 _snake_case = 15_36 _snake_case = 12 _snake_case = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): _snake_case = 10_24 _snake_case = 40_96 _snake_case = 24 _snake_case = 16 # load original model from timm _snake_case = timm.create_model(SCREAMING_SNAKE_CASE__ , pretrained=SCREAMING_SNAKE_CASE__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys _snake_case = timm_model.state_dict() _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # load HuggingFace model _snake_case = DeiTForImageClassificationWithTeacher(SCREAMING_SNAKE_CASE__ ).eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image, prepared by DeiTImageProcessor _snake_case = int( (2_56 / 2_24) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 _snake_case = DeiTImageProcessor(size=SCREAMING_SNAKE_CASE__ , crop_size=config.image_size ) _snake_case = image_processor(images=prepare_img() , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) _snake_case = timm_model(SCREAMING_SNAKE_CASE__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE__ , outputs.logits , atol=1E-3 ) Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--deit_name""", default="""vit_deit_base_distilled_patch16_224""", type=str, help="""Name of the DeiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) __magic_name__ : Optional[int] = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
672
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __magic_name__ : Optional[int] = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ['''pixel_values'''] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BILINEAR , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = size if size is not None else {"shortest_edge": 256} _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) _snake_case = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): _snake_case = get_size_dict(lowerCamelCase ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = size if size is not None else self.size _snake_case = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(lowerCamelCase ) _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] _snake_case = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] _snake_case = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
672
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() __magic_name__ : List[Any] = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' _snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _snake_case = "" else: _snake_case = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) _snake_case = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[ : config.hidden_size, : ] _snake_case = in_proj_bias[: config.hidden_size] _snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case = in_proj_weight[ -config.hidden_size :, : ] _snake_case = in_proj_bias[-config.hidden_size :] def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dct.pop(SCREAMING_SNAKE_CASE__ ) _snake_case = val def snake_case_ ( ): '''simple docstring''' _snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=True ): '''simple docstring''' _snake_case = ViTConfig() # patch_size if model_name[-1] == "8": _snake_case = 8 # set labels if required if not base_model: _snake_case = 10_00 _snake_case = "huggingface/label-files" _snake_case = "imagenet-1k-id2label.json" _snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="dataset" ) , "r" ) ) _snake_case = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _snake_case = 3_84 _snake_case = 15_36 _snake_case = 12 _snake_case = 6 # load original model from torch hub _snake_case = torch.hub.load("facebookresearch/dino:main" , SCREAMING_SNAKE_CASE__ ) original_model.eval() # load state_dict of original model, remove and rename some keys _snake_case = original_model.state_dict() if base_model: remove_classification_head_(SCREAMING_SNAKE_CASE__ ) _snake_case = create_rename_keys(SCREAMING_SNAKE_CASE__ , base_model=SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # load HuggingFace model if base_model: _snake_case = ViTModel(SCREAMING_SNAKE_CASE__ , add_pooling_layer=SCREAMING_SNAKE_CASE__ ).eval() else: _snake_case = ViTForImageClassification(SCREAMING_SNAKE_CASE__ ).eval() model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image, prepared by ViTImageProcessor _snake_case = ViTImageProcessor() _snake_case = image_processor(images=prepare_img() , return_tensors="pt" ) _snake_case = encoding["pixel_values"] _snake_case = model(SCREAMING_SNAKE_CASE__ ) if base_model: _snake_case = original_model(SCREAMING_SNAKE_CASE__ ) assert torch.allclose(SCREAMING_SNAKE_CASE__ , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _snake_case = original_model(SCREAMING_SNAKE_CASE__ ) assert logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE__ , outputs.logits , atol=1E-3 ) Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""dino_vitb16""", type=str, help="""Name of the model trained with DINO you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether to only convert the base model (no projection head weights).""", ) parser.set_defaults(base_model=True) __magic_name__ : Optional[Any] = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
672
'''simple docstring''' import baseaa def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return baseaa.aaadecode(SCREAMING_SNAKE_CASE__ ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
672
1
'''simple docstring''' from collections import defaultdict def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = first_str.lower().strip() _snake_case = second_str.lower().strip() # Remove whitespace _snake_case = first_str.replace(" " , "" ) _snake_case = second_str.replace(" " , "" ) # Strings of different lengths are not anagrams if len(SCREAMING_SNAKE_CASE__ ) != len(SCREAMING_SNAKE_CASE__ ): return False # Default values for count should be 0 _snake_case = defaultdict(SCREAMING_SNAKE_CASE__ ) # For each character in input strings, # increment count in the corresponding for i in range(len(SCREAMING_SNAKE_CASE__ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() __magic_name__ : List[str] = input("""Enter the first string """).strip() __magic_name__ : Optional[int] = input("""Enter the second string """).strip() __magic_name__ : Dict = check_anagrams(input_a, input_b) print(F'{input_a} and {input_b} are {"" if status else "not "}anagrams.')
672
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
672
1
'''simple docstring''' import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __magic_name__ : Optional[Any] = """platform""" import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class __SCREAMING_SNAKE_CASE : '''simple docstring''' UpperCAmelCase__ : Optional[Any] = PegasusConfig UpperCAmelCase__ : Any = {} UpperCAmelCase__ : Optional[int] = '''gelu''' def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=20 , lowerCamelCase=2 , lowerCamelCase=1 , lowerCamelCase=0 , ): _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id def UpperCamelCase( self ): _snake_case = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ).clip(3 , self.vocab_size ) _snake_case = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case = np.concatenate([input_ids, eos_tensor] , axis=1 ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _snake_case = prepare_pegasus_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, inputs_dict def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = 20 _snake_case = model_class_name(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] ) _snake_case , _snake_case = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase , lowerCamelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCamelCase , ) _snake_case = model.decode(lowerCamelCase , lowerCamelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): _snake_case = 20 _snake_case = model_class_name(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] ) _snake_case , _snake_case = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase , lowerCamelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase , decoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCamelCase , decoder_position_ids=lowerCamelCase , ) _snake_case = model.decode(lowerCamelCase , lowerCamelCase , decoder_attention_mask=lowerCamelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , ): '''simple docstring''' if attention_mask is None: _snake_case = np.not_equal(SCREAMING_SNAKE_CASE__ , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: _snake_case = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) UpperCAmelCase__ : Tuple = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Any = False UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Dict = False def UpperCamelCase( self ): _snake_case = FlaxPegasusModelTester(self ) _snake_case = ConfigTester(self , config_class=lowerCamelCase ) def UpperCamelCase( self ): self.config_tester.run_common_tests() def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) _snake_case = model_class(lowerCamelCase ) @jax.jit def encode_jitted(lowerCamelCase , lowerCamelCase=None , **lowerCamelCase ): return model.encode(input_ids=lowerCamelCase , attention_mask=lowerCamelCase ) with self.subTest("JIT Enabled" ): _snake_case = encode_jitted(**lowerCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _snake_case = encode_jitted(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for jitted_output, output in zip(lowerCamelCase , lowerCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCamelCase( self ): _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(lowerCamelCase ) _snake_case = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) _snake_case = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowerCamelCase , lowerCamelCase , lowerCamelCase ): return model.decode( decoder_input_ids=lowerCamelCase , decoder_attention_mask=lowerCamelCase , encoder_outputs=lowerCamelCase , ) with self.subTest("JIT Enabled" ): _snake_case = decode_jitted(**lowerCamelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _snake_case = decode_jitted(**lowerCamelCase ).to_tuple() self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) for jitted_output, output in zip(lowerCamelCase , lowerCamelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCamelCase( self ): for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("google/pegasus-large" , from_pt=lowerCamelCase ) _snake_case = np.ones((1, 1) ) _snake_case = model(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @slow def UpperCamelCase( self ): _snake_case = FlaxPegasusForConditionalGeneration.from_pretrained("google/pegasus-xsum" ) _snake_case = PegasusTokenizer.from_pretrained("google/pegasus-xsum" ) _snake_case = [ " PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.", " The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" ", ] _snake_case = [ "California's largest electricity provider has turned off power to hundreds of thousands of customers.", "Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.", ] _snake_case = tokenizer(lowerCamelCase , return_tensors="np" , truncation=lowerCamelCase , max_length=512 , padding=lowerCamelCase ) _snake_case = model.generate(**lowerCamelCase , num_beams=2 ).sequences _snake_case = tokenizer.batch_decode(lowerCamelCase , skip_special_tokens=lowerCamelCase ) assert tgt_text == decoded
672
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): '''simple docstring''' def UpperCamelCase( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCamelCase , ) def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCamelCase ) def snake_case_ ( ): '''simple docstring''' return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case_ ( ): '''simple docstring''' return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): import apache_beam as beam _snake_case = beam.io.parquetio.WriteToParquet _snake_case = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: _snake_case = partial(lowerCamelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCamelCase( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = DummyBeamDataset(cache_dir=lowerCamelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCamelCase( self ): _snake_case = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _snake_case = NestedBeamDataset(cache_dir=lowerCamelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , F'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) _snake_case = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCamelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCamelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCamelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
672
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: __magic_name__ : Any = None __magic_name__ : Optional[Any] = logging.get_logger(__name__) __magic_name__ : Any = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} __magic_name__ : Tuple = { """vocab_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json""" ), }, } __magic_name__ : int = { """moussaKam/mbarthez""": 1_024, """moussaKam/barthez""": 1_024, """moussaKam/barthez-orangesum-title""": 1_024, } __magic_name__ : Dict = """▁""" class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Any = VOCAB_FILES_NAMES UpperCAmelCase__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Dict = ['''input_ids''', '''attention_mask'''] UpperCAmelCase__ : Optional[Any] = BarthezTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="</s>" , lowerCamelCase="<s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase="<mask>" , **lowerCamelCase , ): # Mask token behave like a normal word, i.e. include the space before it _snake_case = AddedToken(lowerCamelCase , lstrip=lowerCamelCase , rstrip=lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) else mask_token super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , cls_token=lowerCamelCase , pad_token=lowerCamelCase , mask_token=lowerCamelCase , **lowerCamelCase , ) _snake_case = vocab_file _snake_case = False if not self.vocab_file else True def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _snake_case = [self.cls_token_id] _snake_case = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): _snake_case = [self.sep_token_id] _snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase( self , lowerCamelCase , lowerCamelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCamelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _snake_case = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ): copyfile(self.vocab_file , lowerCamelCase ) return (out_vocab_file,)
672
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __magic_name__ : Optional[int] = False class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _snake_case = torch.manual_seed(0 ) _snake_case = pipe( image=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _snake_case = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _snake_case = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
672
1
'''simple docstring''' import qiskit def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register _snake_case = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator _snake_case = qiskit.execute(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , shots=10_00 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": print(F'Total count for various states are: {single_qubit_measure(1, 1)}')
672
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = text_path elif issubclass(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _snake_case = [text_path] _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_dataset(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=("train",) ): '''simple docstring''' assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for split in splits: _snake_case = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _snake_case = TextDatasetReader({"train": text_path} , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _snake_case = {"text": "string"} _snake_case = features.copy() if features else default_expected_features _snake_case = ( Features({feature: Value(SCREAMING_SNAKE_CASE__ ) for feature, dtype in features.items()} ) if features is not None else None ) _snake_case = TextDatasetReader({"train": text_path} , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if split: _snake_case = {split: text_path} else: _snake_case = "train" _snake_case = {"train": text_path, "test": text_path} _snake_case = tmp_path / "cache" _snake_case = {"text": "string"} _snake_case = TextDatasetReader(SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ ).read() _check_text_datasetdict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
672
1
'''simple docstring''' import colorsys from PIL import Image # type: ignore def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = x _snake_case = y for step in range(SCREAMING_SNAKE_CASE__ ): # noqa: B007 _snake_case = a * a - b * b + x _snake_case = 2 * a * b + y _snake_case = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(SCREAMING_SNAKE_CASE__ , 1 , 1 ) ) def snake_case_ ( SCREAMING_SNAKE_CASE__ = 8_00 , SCREAMING_SNAKE_CASE__ = 6_00 , SCREAMING_SNAKE_CASE__ = -0.6 , SCREAMING_SNAKE_CASE__ = 0 , SCREAMING_SNAKE_CASE__ = 3.2 , SCREAMING_SNAKE_CASE__ = 50 , SCREAMING_SNAKE_CASE__ = True , ): '''simple docstring''' _snake_case = Image.new("RGB" , (image_width, image_height) ) _snake_case = img.load() # loop through the image-coordinates for image_x in range(SCREAMING_SNAKE_CASE__ ): for image_y in range(SCREAMING_SNAKE_CASE__ ): # determine the figure-coordinates based on the image-coordinates _snake_case = figure_width / image_width * image_height _snake_case = figure_center_x + (image_x / image_width - 0.5) * figure_width _snake_case = figure_center_y + (image_y / image_height - 0.5) * figure_height _snake_case = get_distance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _snake_case = get_color_coded_rgb(SCREAMING_SNAKE_CASE__ ) else: _snake_case = get_black_and_white_rgb(SCREAMING_SNAKE_CASE__ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure __magic_name__ : Tuple = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
672
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ : Any = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Dict = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __magic_name__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
672
1
'''simple docstring''' import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = VideoToVideoSDPipeline UpperCAmelCase__ : Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'''video'''} ) - {'''image''', '''width''', '''height'''} UpperCAmelCase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''video'''} ) - {'''image'''} UpperCAmelCase__ : str = PipelineTesterMixin.required_optional_params - {'''latents'''} UpperCAmelCase__ : Dict = False # No `output_type`. UpperCAmelCase__ : Optional[Any] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def UpperCamelCase( self ): torch.manual_seed(0 ) _snake_case = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCamelCase , set_alpha_to_one=lowerCamelCase , ) torch.manual_seed(0 ) _snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="gelu" , projection_dim=512 , ) _snake_case = CLIPTextModel(lowerCamelCase ) _snake_case = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def UpperCamelCase( self , lowerCamelCase , lowerCamelCase=0 ): # 3 frames _snake_case = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) if str(lowerCamelCase ).startswith("mps" ): _snake_case = torch.manual_seed(lowerCamelCase ) else: _snake_case = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) _snake_case = { "prompt": "A painting of a squirrel eating a burger", "video": video, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def UpperCamelCase( self ): _snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case = self.get_dummy_components() _snake_case = VideoToVideoSDPipeline(**lowerCamelCase ) _snake_case = sd_pipe.to(lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase ) _snake_case = self.get_dummy_inputs(lowerCamelCase ) _snake_case = "np" _snake_case = sd_pipe(**lowerCamelCase ).frames _snake_case = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _snake_case = np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase , expected_max_diff=5e-3 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def UpperCamelCase( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def UpperCamelCase( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): return super().test_progress_bar() @slow @skip_mps class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): _snake_case = VideoToVideoSDPipeline.from_pretrained("cerspense/zeroscope_v2_XL" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames _snake_case = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case = torch.randn((1, 10, 3, 1_024, 576) , generator=lowerCamelCase ) _snake_case = video.to("cuda" ) _snake_case = "Spiderman is surfing" _snake_case = pipe(lowerCamelCase , video=lowerCamelCase , generator=lowerCamelCase , num_inference_steps=3 , output_type="pt" ).frames _snake_case = np.array([-1.0458984, -1.1279297, -0.9663086, -0.91503906, -0.75097656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
672
'''simple docstring''' import math def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return math.pow(SCREAMING_SNAKE_CASE__ , 2 ) - a def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return 2 * x def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 2.0 while start <= a: _snake_case = math.pow(SCREAMING_SNAKE_CASE__ , 2 ) return start def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 99_99 , SCREAMING_SNAKE_CASE__ = 0.00000000000001 ): '''simple docstring''' if a < 0: raise ValueError("math domain error" ) _snake_case = get_initial_point(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): _snake_case = value _snake_case = value - fx(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) / fx_derivative(SCREAMING_SNAKE_CASE__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
672
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=18 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , ): _snake_case = size if size is not None else {"shortest_edge": 20} _snake_case = crop_size if crop_size is not None else {"height": 18, "width": 18} _snake_case = parent _snake_case = batch_size _snake_case = num_channels _snake_case = image_size _snake_case = min_resolution _snake_case = max_resolution _snake_case = do_resize _snake_case = size _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_flip_channel_order def UpperCamelCase( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( __UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = MobileViTImageProcessor if is_vision_available() else None def UpperCamelCase( self ): _snake_case = MobileViTImageProcessingTester(self ) @property def UpperCamelCase( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase( self ): _snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_flip_channel_order" ) ) def UpperCamelCase( self ): _snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 20} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) _snake_case = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _snake_case = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def UpperCamelCase( self ): # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _snake_case = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def UpperCamelCase( self ): # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched _snake_case = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
672
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[int] = logging.get_logger(__name__) __magic_name__ : Optional[int] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Tuple = '''git_vision_model''' def __init__( self , lowerCamelCase=768 , lowerCamelCase=3_072 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=16 , lowerCamelCase="quick_gelu" , lowerCamelCase=1e-5 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) _snake_case = hidden_size _snake_case = intermediate_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = num_channels _snake_case = patch_size _snake_case = image_size _snake_case = initializer_range _snake_case = attention_dropout _snake_case = layer_norm_eps _snake_case = hidden_act @classmethod def UpperCamelCase( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(lowerCamelCase ) _snake_case , _snake_case = cls.get_config_dict(lowerCamelCase , **lowerCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": _snake_case = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase , **lowerCamelCase ) class __SCREAMING_SNAKE_CASE ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : int = '''git''' def __init__( self , lowerCamelCase=None , lowerCamelCase=30_522 , lowerCamelCase=768 , lowerCamelCase=6 , lowerCamelCase=12 , lowerCamelCase=3_072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1_024 , lowerCamelCase=0.02 , lowerCamelCase=1e-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=101 , lowerCamelCase=102 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , pad_token_id=lowerCamelCase , **lowerCamelCase ) if vision_config is None: _snake_case = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) _snake_case = GitVisionConfig(**lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache _snake_case = tie_word_embeddings _snake_case = num_image_with_embedding _snake_case = bos_token_id _snake_case = eos_token_id def UpperCamelCase( self ): _snake_case = copy.deepcopy(self.__dict__ ) _snake_case = self.vision_config.to_dict() _snake_case = self.__class__.model_type return output
672
1
'''simple docstring''' import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration __magic_name__ : int = 500_000 __magic_name__ , __magic_name__ : int = os.path.split(__file__) __magic_name__ : Optional[int] = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def snake_case_ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dataset.map(**SCREAMING_SNAKE_CASE__ ) @get_duration def snake_case_ ( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = dataset.filter(**SCREAMING_SNAKE_CASE__ ) def snake_case_ ( ): '''simple docstring''' _snake_case = {"num examples": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = datasets.Features({"text": datasets.Value("string" ), "numbers": datasets.Value("float32" )} ) _snake_case = generate_example_dataset( os.path.join(SCREAMING_SNAKE_CASE__ , "dataset.arrow" ) , SCREAMING_SNAKE_CASE__ , num_examples=SCREAMING_SNAKE_CASE__ ) _snake_case = transformers.AutoTokenizer.from_pretrained("bert-base-cased" , use_fast=SCREAMING_SNAKE_CASE__ ) def tokenize(SCREAMING_SNAKE_CASE__ ): return tokenizer(examples["text"] ) _snake_case = map(SCREAMING_SNAKE_CASE__ ) _snake_case = map(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) _snake_case = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type="numpy" ): _snake_case = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type="pandas" ): _snake_case = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type="torch" , columns="numbers" ): _snake_case = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) with dataset.formatted_as(type="tensorflow" , columns="numbers" ): _snake_case = map(SCREAMING_SNAKE_CASE__ , function=lambda SCREAMING_SNAKE_CASE__ : None , batched=SCREAMING_SNAKE_CASE__ ) _snake_case = map(SCREAMING_SNAKE_CASE__ , function=SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) _snake_case = filter(SCREAMING_SNAKE_CASE__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(SCREAMING_SNAKE_CASE__ , "wb" ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
672
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __magic_name__ : Dict = logging.get_logger(__name__) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if isinstance(SCREAMING_SNAKE_CASE__ , np.ndarray ): return list(tensor.shape ) _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) if tensor.shape == tf.TensorShape(SCREAMING_SNAKE_CASE__ ): return dynamic _snake_case = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(SCREAMING_SNAKE_CASE__ )] def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9 , axis=SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1E-5 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise NotImplementedError("Only 1D weight and bias tensors are supported for now, with only a single axis." ) # Get mean and variance on the axis to be normalized _snake_case , _snake_case = tf.nn.moments(SCREAMING_SNAKE_CASE__ , axes=[axis] , keepdims=SCREAMING_SNAKE_CASE__ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis _snake_case = [1] * inputs.shape.rank _snake_case = shape_list(SCREAMING_SNAKE_CASE__ )[axis] _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _snake_case = tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Compute layer normalization using the batch_normalization # function. _snake_case = tf.nn.batch_normalization( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , offset=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , variance_epsilon=SCREAMING_SNAKE_CASE__ , ) return outputs def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input _snake_case = tf.shape(SCREAMING_SNAKE_CASE__ ) _snake_case = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) _snake_case = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ): _snake_case = tf.convert_to_tensor(SCREAMING_SNAKE_CASE__ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: _snake_case = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: _snake_case = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) _snake_case = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( SCREAMING_SNAKE_CASE__ , tf.cast(SCREAMING_SNAKE_CASE__ , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(SCREAMING_SNAKE_CASE__ )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = 6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. _snake_case = [x for x in data if len(SCREAMING_SNAKE_CASE__ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( "The following attributes cannot be saved to HDF5 file because " f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) _snake_case = np.asarray(SCREAMING_SNAKE_CASE__ ) _snake_case = 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 _snake_case = np.array_split(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(SCREAMING_SNAKE_CASE__ ): _snake_case = chunk_data else: _snake_case = data def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if name in group.attrs: _snake_case = [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs[name]] else: _snake_case = [] _snake_case = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode("utf8" ) if hasattr(SCREAMING_SNAKE_CASE__ , "decode" ) else n for n in group.attrs["%s%d" % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _expand_single_ad_tensor(SCREAMING_SNAKE_CASE__ ): if isinstance(SCREAMING_SNAKE_CASE__ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(SCREAMING_SNAKE_CASE__ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , SCREAMING_SNAKE_CASE__ )
672
1
'''simple docstring''' def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return base * power(SCREAMING_SNAKE_CASE__ , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print("""Raise base to the power of exponent using recursion...""") __magic_name__ : str = int(input("""Enter the base: """).strip()) __magic_name__ : Optional[int] = int(input("""Enter the exponent: """).strip()) __magic_name__ : List[str] = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents __magic_name__ : List[Any] = 1 / result print(F'{base} to the power of {exponent} is {result}')
672
'''simple docstring''' __magic_name__ : int = """Alexander Joslin""" import operator as op from .stack import Stack def snake_case_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case = Stack() _snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE__ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE__ ) elif i == ")": # RULE 4 _snake_case = operator_stack.peek() operator_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operand_stack.peek() operand_stack.pop() _snake_case = operators[opr](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) operand_stack.push(SCREAMING_SNAKE_CASE__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __magic_name__ : List[str] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
672
1