code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( '''--original_config_file''', default=None, type=str, help='''The YAML config file corresponding to the original architecture.''', ) parser.add_argument( '''--num_in_channels''', default=None, type=int, help='''The number of input channels. If `None` number of input channels will be automatically inferred.''', ) parser.add_argument( '''--scheduler_type''', default='''pndm''', type=str, help='''Type of scheduler to use. Should be one of [\'pndm\', \'lms\', \'ddim\', \'euler\', \'euler-ancestral\', \'dpm\']''', ) parser.add_argument( '''--pipeline_type''', default=None, type=str, help=( '''The pipeline type. One of \'FrozenOpenCLIPEmbedder\', \'FrozenCLIPEmbedder\', \'PaintByExample\'''' '''. If `None` pipeline will be automatically inferred.''' ), ) parser.add_argument( '''--image_size''', default=None, type=int, help=( '''The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2''' ''' Base. Use 768 for Stable Diffusion v2.''' ), ) parser.add_argument( '''--prediction_type''', default=None, type=str, help=( '''The prediction type that the model was trained on. Use \'epsilon\' for Stable Diffusion v1.X and Stable''' ''' Diffusion v2 Base. Use \'v_prediction\' for Stable Diffusion v2.''' ), ) parser.add_argument( '''--extract_ema''', action='''store_true''', help=( '''Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights''' ''' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield''' ''' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.''' ), ) parser.add_argument( '''--upcast_attention''', action='''store_true''', help=( '''Whether the attention computation should always be upcasted. This is necessary when running stable''' ''' diffusion 2.1.''' ), ) parser.add_argument( '''--from_safetensors''', action='''store_true''', help='''If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.''', ) parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') parser.add_argument( '''--stable_unclip''', type=str, default=None, required=False, help='''Set if this is a stable unCLIP model. One of \'txt2img\' or \'img2img\'.''', ) parser.add_argument( '''--stable_unclip_prior''', type=str, default=None, required=False, help='''Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.''', ) parser.add_argument( '''--clip_stats_path''', type=str, help='''Path to the clip stats file. Only required if the stable unclip model\'s config specifies `model.params.noise_aug_config.params.clip_stats_path`.''', required=False, ) parser.add_argument( '''--controlnet''', action='''store_true''', default=None, help='''Set flag if this is a controlnet checkpoint.''' ) parser.add_argument('''--half''', action='''store_true''', help='''Save weights in half precision.''') parser.add_argument( '''--vae_path''', type=str, default=None, required=False, help='''Set to a path, hub id to an already converted vae to not convert it again.''', ) a : Union[str, Any] = parser.parse_args() a : List[Any] = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" import random def _SCREAMING_SNAKE_CASE ( _lowercase : list , _lowercase : Optional[Any] ) ->tuple: '''simple docstring''' a, a, a : List[Any] = [], [], [] for element in data: if element < pivot: less.append(_lowercase ) elif element > pivot: greater.append(_lowercase ) else: equal.append(_lowercase ) return less, equal, greater def _SCREAMING_SNAKE_CASE ( _lowercase : list , _lowercase : int ) ->Dict: '''simple docstring''' if index >= len(_lowercase ) or index < 0: return None a : Optional[int] = items[random.randint(0 , len(_lowercase ) - 1 )] a : Dict = 0 a, a, a : str = _partition(_lowercase , _lowercase ) a : List[Any] = len(_lowercase ) a : int = len(_lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(_lowercase , _lowercase ) # must be in larger else: return quick_select(_lowercase , index - (m + count) )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize("dataset_size" , [None, 400 * 2**20, 600 * 2**20] ) @pytest.mark.parametrize("input_in_memory_max_size" , ["default", 0, 100 * 2**20, 900 * 2**20] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Any , _lowercase : Dict ) ->Optional[int]: '''simple docstring''' if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , "IN_MEMORY_MAX_SIZE" , _lowercase ) a : Optional[int] = datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: a : List[str] = dataset_size < in_memory_max_size else: a : str = False a : Optional[int] = is_small_dataset(_lowercase ) assert result == expected
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Dict = { '''microsoft/git-base''': '''https://huggingface.co/microsoft/git-base/resolve/main/config.json''', } class __UpperCamelCase ( a__ ): lowerCamelCase : int ="""git_vision_model""" def __init__( self , lowerCAmelCase__=768 , lowerCAmelCase__=3072 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=3 , lowerCAmelCase__=224 , lowerCAmelCase__=16 , lowerCAmelCase__="quick_gelu" , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.02 , **lowerCAmelCase__ , ) -> str: super().__init__(**lowerCAmelCase__ ) a : Tuple = hidden_size a : List[Any] = intermediate_size a : List[Any] = num_hidden_layers a : Tuple = num_attention_heads a : str = num_channels a : Union[str, Any] = patch_size a : List[Any] = image_size a : Optional[int] = initializer_range a : Tuple = attention_dropout a : Union[str, Any] = layer_norm_eps a : List[Any] = hidden_act @classmethod def __a ( cls , lowerCAmelCase__ , **lowerCAmelCase__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(lowerCAmelCase__ ) a, a : List[str] = cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": a : Dict = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class __UpperCamelCase ( a__ ): lowerCamelCase : List[Any] ="""git""" def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=3_0522 , lowerCAmelCase__=768 , lowerCAmelCase__=6 , lowerCAmelCase__=12 , lowerCAmelCase__=3072 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1024 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__="absolute" , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=101 , lowerCAmelCase__=102 , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> List[Any]: super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) if vision_config is None: a : List[str] = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) a : Optional[int] = GitVisionConfig(**lowerCAmelCase__ ) a : str = vocab_size a : Union[str, Any] = hidden_size a : str = num_hidden_layers a : List[str] = num_attention_heads a : Any = hidden_act a : Tuple = intermediate_size a : List[Any] = hidden_dropout_prob a : Tuple = attention_probs_dropout_prob a : str = max_position_embeddings a : int = initializer_range a : Optional[int] = layer_norm_eps a : int = position_embedding_type a : Optional[Any] = use_cache a : Any = tie_word_embeddings a : List[Any] = num_image_with_embedding a : int = bos_token_id a : str = eos_token_id def __a ( self ) -> str: a : int = copy.deepcopy(self.__dict__ ) a : Optional[int] = self.vision_config.to_dict() a : Optional[Any] = self.__class__.model_type return output
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# a : Tuple = [ # (stable-diffusion, HF Diffusers) ('''time_embed.0.weight''', '''time_embedding.linear_1.weight'''), ('''time_embed.0.bias''', '''time_embedding.linear_1.bias'''), ('''time_embed.2.weight''', '''time_embedding.linear_2.weight'''), ('''time_embed.2.bias''', '''time_embedding.linear_2.bias'''), ('''input_blocks.0.0.weight''', '''conv_in.weight'''), ('''input_blocks.0.0.bias''', '''conv_in.bias'''), ('''out.0.weight''', '''conv_norm_out.weight'''), ('''out.0.bias''', '''conv_norm_out.bias'''), ('''out.2.weight''', '''conv_out.weight'''), ('''out.2.bias''', '''conv_out.bias'''), ] a : Tuple = [ # (stable-diffusion, HF Diffusers) ('''in_layers.0''', '''norm1'''), ('''in_layers.2''', '''conv1'''), ('''out_layers.0''', '''norm2'''), ('''out_layers.3''', '''conv2'''), ('''emb_layers.1''', '''time_emb_proj'''), ('''skip_connection''', '''conv_shortcut'''), ] a : List[str] = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks a : Tuple = F'''down_blocks.{i}.resnets.{j}.''' a : Optional[int] = F'''input_blocks.{3*i + j + 1}.0.''' unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 a : Dict = F'''down_blocks.{i}.attentions.{j}.''' a : int = F'''input_blocks.{3*i + j + 1}.1.''' unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks a : Dict = F'''up_blocks.{i}.resnets.{j}.''' a : int = F'''output_blocks.{3*i + j}.0.''' unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 a : Any = F'''up_blocks.{i}.attentions.{j}.''' a : Optional[int] = F'''output_blocks.{3*i + j}.1.''' unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 a : Dict = F'''down_blocks.{i}.downsamplers.0.conv.''' a : int = F'''input_blocks.{3*(i+1)}.0.op.''' unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 a : Tuple = F'''up_blocks.{i}.upsamplers.0.''' a : Optional[Any] = F'''output_blocks.{3*i + 2}.{1 if i == 0 else 2}.''' unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) a : Optional[Any] = '''mid_block.attentions.0.''' a : Optional[int] = '''middle_block.1.''' unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): a : List[str] = F'''mid_block.resnets.{j}.''' a : Optional[Any] = F'''middle_block.{2*j}.''' unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->Any: '''simple docstring''' a : Optional[Any] = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: a : str = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: a : str = v.replace(_lowercase , _lowercase ) a : Tuple = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: a : Any = v.replace(_lowercase , _lowercase ) a : Any = v a : Any = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# a : Any = [ # (stable-diffusion, HF Diffusers) ('''nin_shortcut''', '''conv_shortcut'''), ('''norm_out''', '''conv_norm_out'''), ('''mid.attn_1.''', '''mid_block.attentions.0.'''), ] for i in range(4): # down_blocks have two resnets for j in range(2): a : List[str] = F'''encoder.down_blocks.{i}.resnets.{j}.''' a : Dict = F'''encoder.down.{i}.block.{j}.''' vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: a : int = F'''down_blocks.{i}.downsamplers.0.''' a : List[Any] = F'''down.{i}.downsample.''' vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) a : Tuple = F'''up_blocks.{i}.upsamplers.0.''' a : List[str] = F'''up.{3-i}.upsample.''' vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): a : Union[str, Any] = F'''decoder.up_blocks.{i}.resnets.{j}.''' a : Optional[Any] = F'''decoder.up.{3-i}.block.{j}.''' vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): a : Dict = F'''mid_block.resnets.{i}.''' a : Tuple = F'''mid.block_{i+1}.''' vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) a : Optional[int] = [ # (stable-diffusion, HF Diffusers) ('''norm.''', '''group_norm.'''), ('''q.''', '''query.'''), ('''k.''', '''key.'''), ('''v.''', '''value.'''), ('''proj_out.''', '''proj_attn.'''), ] def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->int: '''simple docstring''' return w.reshape(*w.shape , 1 , 1 ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Optional[int]: '''simple docstring''' a : Optional[int] = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: a : Tuple = v.replace(_lowercase , _lowercase ) a : int = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: a : int = v.replace(_lowercase , _lowercase ) a : int = v a : Dict = {v: vae_state_dict[k] for k, v in mapping.items()} a : int = ["q", "k", "v", "proj_out"] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F"""mid.attn_1.{weight_name}.weight""" in k: print(F"""Reshaping {k} for SD format""" ) a : Optional[int] = reshape_weight_for_sd(_lowercase ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# a : Optional[Any] = [ # (stable-diffusion, HF Diffusers) ('''resblocks.''', '''text_model.encoder.layers.'''), ('''ln_1''', '''layer_norm1'''), ('''ln_2''', '''layer_norm2'''), ('''.c_fc.''', '''.fc1.'''), ('''.c_proj.''', '''.fc2.'''), ('''.attn''', '''.self_attn'''), ('''ln_final.''', '''transformer.text_model.final_layer_norm.'''), ('''token_embedding.weight''', '''transformer.text_model.embeddings.token_embedding.weight'''), ('''positional_embedding''', '''transformer.text_model.embeddings.position_embedding.weight'''), ] a : Dict = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} a : str = re.compile('''|'''.join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp a : Optional[int] = {'''q''': 0, '''k''': 1, '''v''': 2} def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->Union[str, Any]: '''simple docstring''' a : int = {} a : Optional[Any] = {} a : Dict = {} for k, v in text_enc_dict.items(): if ( k.endswith(".self_attn.q_proj.weight" ) or k.endswith(".self_attn.k_proj.weight" ) or k.endswith(".self_attn.v_proj.weight" ) ): a : Dict = k[: -len(".q_proj.weight" )] a : str = k[-len("q_proj.weight" )] if k_pre not in capture_qkv_weight: a : List[Any] = [None, None, None] a : Union[str, Any] = v continue if ( k.endswith(".self_attn.q_proj.bias" ) or k.endswith(".self_attn.k_proj.bias" ) or k.endswith(".self_attn.v_proj.bias" ) ): a : str = k[: -len(".q_proj.bias" )] a : Tuple = k[-len("q_proj.bias" )] if k_pre not in capture_qkv_bias: a : List[Any] = [None, None, None] a : str = v continue a : str = textenc_pattern.sub(lambda _lowercase : protected[re.escape(m.group(0 ) )] , _lowercase ) a : Any = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception("CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing" ) a : str = textenc_pattern.sub(lambda _lowercase : protected[re.escape(m.group(0 ) )] , _lowercase ) a : Dict = torch.cat(_lowercase ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception("CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing" ) a : Optional[Any] = textenc_pattern.sub(lambda _lowercase : protected[re.escape(m.group(0 ) )] , _lowercase ) a : str = torch.cat(_lowercase ) return new_state_dict def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] ) ->Optional[int]: '''simple docstring''' return text_enc_dict if __name__ == "__main__": a : List[str] = argparse.ArgumentParser() parser.add_argument('''--model_path''', default=None, type=str, required=True, help='''Path to the model to convert.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--half''', action='''store_true''', help='''Save weights in half precision.''') parser.add_argument( '''--use_safetensors''', action='''store_true''', help='''Save weights use safetensors, default is ckpt.''' ) a : List[str] = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors a : Any = osp.join(args.model_path, '''unet''', '''diffusion_pytorch_model.safetensors''') a : Union[str, Any] = osp.join(args.model_path, '''vae''', '''diffusion_pytorch_model.safetensors''') a : Dict = osp.join(args.model_path, '''text_encoder''', '''model.safetensors''') # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): a : Dict = load_file(unet_path, device='''cpu''') else: a : Optional[Any] = osp.join(args.model_path, '''unet''', '''diffusion_pytorch_model.bin''') a : int = torch.load(unet_path, map_location='''cpu''') if osp.exists(vae_path): a : Dict = load_file(vae_path, device='''cpu''') else: a : Any = osp.join(args.model_path, '''vae''', '''diffusion_pytorch_model.bin''') a : str = torch.load(vae_path, map_location='''cpu''') if osp.exists(text_enc_path): a : str = load_file(text_enc_path, device='''cpu''') else: a : Union[str, Any] = osp.join(args.model_path, '''text_encoder''', '''pytorch_model.bin''') a : Dict = torch.load(text_enc_path, map_location='''cpu''') # Convert the UNet model a : Any = convert_unet_state_dict(unet_state_dict) a : List[Any] = {'''model.diffusion_model.''' + k: v for k, v in unet_state_dict.items()} # Convert the VAE model a : Optional[int] = convert_vae_state_dict(vae_state_dict) a : Tuple = {'''first_stage_model.''' + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper a : Optional[int] = '''text_model.encoder.layers.22.layer_norm2.bias''' in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm a : int = {'''transformer.''' + k: v for k, v in text_enc_dict.items()} a : List[Any] = convert_text_enc_state_dict_vaa(text_enc_dict) a : Optional[int] = {'''cond_stage_model.model.''' + k: v for k, v in text_enc_dict.items()} else: a : str = convert_text_enc_state_dict(text_enc_dict) a : List[Any] = {'''cond_stage_model.transformer.''' + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint a : Any = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: a : Optional[Any] = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: a : str = {'''state_dict''': state_dict} torch.save(state_dict, args.checkpoint_path)
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : str , _lowercase : Optional[str] = None ) ->str: '''simple docstring''' if version.parse(hfh.__version__ ).release < version.parse("0.11.0" ).release: # old versions of hfh don't url-encode the file path a : List[Any] = quote(_lowercase ) return hfh.hf_hub_url(_lowercase , _lowercase , repo_type="dataset" , revision=_lowercase )
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" import operator as op def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->Any: '''simple docstring''' a : str = [] a : List[str] = lambda _lowercase , _lowercase : int(x / y ) # noqa: E731 integer division operation a : Union[str, Any] = { "^": op.pow, "*": op.mul, "/": div, "+": op.add, "-": op.sub, } # operators & their respective operation # print table header print("Symbol".center(8 ) , "Action".center(12 ) , "Stack" , sep=" | " ) print("-" * (30 + len(_lowercase )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(_lowercase ) # append x to stack # output in tabular format print(x.rjust(8 ) , ("push(" + x + ")").ljust(12 ) , ",".join(_lowercase ) , sep=" | " ) else: a : Optional[Any] = stack.pop() # pop stack # output in tabular format print("".rjust(8 ) , ("pop(" + b + ")").ljust(12 ) , ",".join(_lowercase ) , sep=" | " ) a : List[str] = stack.pop() # pop stack # output in tabular format print("".rjust(8 ) , ("pop(" + a + ")").ljust(12 ) , ",".join(_lowercase ) , sep=" | " ) stack.append( str(opr[x](int(_lowercase ) , int(_lowercase ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ("push(" + a + x + b + ")").ljust(12 ) , ",".join(_lowercase ) , sep=" | " , ) return int(stack[0] ) if __name__ == "__main__": a : Optional[Any] = input('''\n\nEnter a Postfix Equation (space separated) = ''').split(''' ''') print('''\n\tResult = ''', solve(Postfix))
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: a : int = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=3 , lowerCAmelCase__=18 , lowerCAmelCase__=30 , lowerCAmelCase__=400 , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=None , ) -> Optional[Any]: a : Dict = size if size is not None else {"height": 20, "width": 20} a : List[str] = parent a : Dict = batch_size a : Dict = num_channels a : int = image_size a : Any = min_resolution a : Tuple = max_resolution a : Optional[Any] = size a : Any = do_normalize a : Tuple = do_convert_rgb a : Union[str, Any] = [512, 1024, 2048, 4096] a : List[str] = patch_size if patch_size is not None else {"height": 16, "width": 16} def __a ( self ) -> str: return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def __a ( self ) -> Optional[Any]: a : List[Any] = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" a : Tuple = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def __a ( self ) -> Optional[int]: a : List[str] = PixaStructImageProcessingTester(self ) @property def __a ( self ) -> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def __a ( self ) -> List[Any]: a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_normalize" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_convert_rgb" ) ) def __a ( self ) -> List[Any]: a : Any = self.image_processor_tester.prepare_dummy_image() a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) a : Union[str, Any] = 2048 a : Dict = image_processor(lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.0_606 ) , atol=1E-3 , rtol=1E-3 ) ) def __a ( self ) -> Dict: # Initialize image_processor a : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input a : List[Any] = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input a : Tuple = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched a : Optional[int] = image_processor( lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __a ( self ) -> Optional[int]: # Initialize image_processor a : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input a : List[str] = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 a : Any = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(lowerCAmelCase__ ): a : Optional[Any] = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches a : List[Any] = "Hello" a : Dict = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ , header_text=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched a : Union[str, Any] = image_processor( lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ , header_text=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __a ( self ) -> Any: # Initialize image_processor a : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors a : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) a : Union[str, Any] = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input a : int = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched a : Union[str, Any] = image_processor( lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def __a ( self ) -> Optional[Any]: # Initialize image_processor a : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors a : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input a : Dict = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input a : List[str] = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched a : List[Any] = image_processor( lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def __a ( self ) -> str: a : int = PixaStructImageProcessingTester(self , num_channels=4 ) a : List[Any] = 3 @property def __a ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def __a ( self ) -> List[str]: a : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_normalize" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_convert_rgb" ) ) def __a ( self ) -> List[str]: # Initialize image_processor a : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input a : int = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input a : Dict = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched a : Optional[Any] = image_processor( lowerCAmelCase__ , return_tensors="pt" , max_patches=lowerCAmelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import math def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->list[int]: '''simple docstring''' a : int = [] a : str = 2 a : Dict = int(math.sqrt(_lowercase ) ) # Size of every segment a : Dict = [True] * (end + 1) a : List[str] = [] while start <= end: if temp[start] is True: in_prime.append(_lowercase ) for i in range(start * start , end + 1 , _lowercase ): a : Any = False start += 1 prime += in_prime a : Optional[Any] = end + 1 a : List[str] = min(2 * end , _lowercase ) while low <= n: a : List[Any] = [True] * (high - low + 1) for each in in_prime: a : List[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(_lowercase , high + 1 , _lowercase ): a : str = False for j in range(len(_lowercase ) ): if temp[j] is True: prime.append(j + low ) a : Optional[int] = high + 1 a : Dict = min(high + end , _lowercase ) return prime print(sieve(10**6))
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=64 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ) -> Optional[Any]: a : str = parent a : Union[str, Any] = batch_size a : Dict = seq_length a : Optional[Any] = is_training a : Optional[int] = use_input_mask a : str = use_token_type_ids a : Optional[int] = use_labels a : Dict = vocab_size a : Dict = hidden_size a : int = num_hidden_layers a : List[str] = num_attention_heads a : Tuple = intermediate_size a : List[Any] = hidden_act a : Optional[int] = hidden_dropout_prob a : Union[str, Any] = attention_probs_dropout_prob a : str = max_position_embeddings a : Optional[int] = type_vocab_size a : Optional[Any] = type_sequence_label_size a : Optional[int] = initializer_range a : List[str] = num_labels a : Any = num_choices a : Dict = scope a : Any = vocab_size - 1 def __a ( self ) -> Any: a : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : Optional[Any] = None if self.use_input_mask: a : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) a : int = None if self.use_labels: a : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a : Any = self.get_config() return config, input_ids, input_mask, token_labels def __a ( self ) -> Any: return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def __a ( self ) -> int: a, a, a, a : Optional[int] = self.prepare_config_and_inputs() a : Any = True return config, input_ids, input_mask, token_labels def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Dict = GPTNeoXModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Any = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) a : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : int = True a : List[Any] = GPTNeoXModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[int] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: a : List[str] = GPTNeoXForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : int = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: a : str = self.num_labels a : List[str] = GPTNeoXForQuestionAnswering(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Union[str, Any] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : List[str] = self.num_labels a : Dict = GPTNeoXForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : List[str] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: a : Any = self.num_labels a : Tuple = GPTNeoXForTokenClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : List[str] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Tuple = True a : int = GPTNeoXForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() # first forward pass a : Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) a : int = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids a : str = ids_tensor((self.batch_size, 3) , config.vocab_size ) a : Optional[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and a : str = torch.cat([input_ids, next_tokens] , dim=-1 ) a : Dict = torch.cat([input_mask, next_mask] , dim=-1 ) a : Union[str, Any] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) a : int = output_from_no_past["hidden_states"][0] a : Union[str, Any] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] # select random slice a : Optional[int] = ids_tensor((1,) , output_from_past.shape[-1] ).item() a : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx].detach() a : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Tuple: a : List[str] = self.prepare_config_and_inputs() a, a, a, a : Any = config_and_inputs a : List[str] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( a__ , a__ , a__ , unittest.TestCase ): lowerCamelCase : int =( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase : str =(GPTNeoXForCausalLM,) if is_torch_available() else () lowerCamelCase : int =( { """feature-extraction""": GPTNeoXModel, """question-answering""": GPTNeoXForQuestionAnswering, """text-classification""": GPTNeoXForSequenceClassification, """text-generation""": GPTNeoXForCausalLM, """token-classification""": GPTNeoXForTokenClassification, """zero-shot""": GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase : Union[str, Any] =False lowerCamelCase : Optional[Any] =False lowerCamelCase : Union[str, Any] =False lowerCamelCase : str =False def __a ( self ) -> Optional[int]: a : Tuple = GPTNeoXModelTester(self ) a : Dict = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=64 , num_attention_heads=8 ) def __a ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def __a ( self ) -> Union[str, Any]: a, a, a, a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> int: a, a, a, a : str = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Optional[int]: # This regression test was failing with PyTorch < 1.3 a, a, a, a : str = self.model_tester.prepare_config_and_inputs_for_decoder() a : Optional[int] = None self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a, a, a, a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCAmelCase__ ) def __a ( self ) -> str: a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase__ ) def __a ( self ) -> str: a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase__ ) @unittest.skip(reason="Feed forward chunking is not implemented" ) def __a ( self ) -> Union[str, Any]: pass @parameterized.expand([("linear",), ("dynamic",)] ) def __a ( self , lowerCAmelCase__ ) -> str: a, a : Tuple = self.model_tester.prepare_config_and_inputs_for_common() a : Union[str, Any] = ids_tensor([1, 10] , config.vocab_size ) a : Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights a : List[Any] = GPTNeoXModel(lowerCAmelCase__ ) original_model.to(lowerCAmelCase__ ) original_model.eval() a : Tuple = original_model(lowerCAmelCase__ ).last_hidden_state a : Dict = original_model(lowerCAmelCase__ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights a : List[Any] = {"type": scaling_type, "factor": 10.0} a : List[Any] = GPTNeoXModel(lowerCAmelCase__ ) scaled_model.to(lowerCAmelCase__ ) scaled_model.eval() a : str = scaled_model(lowerCAmelCase__ ).last_hidden_state a : Optional[int] = scaled_model(lowerCAmelCase__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5 ) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> int: a : Optional[Any] = AutoTokenizer.from_pretrained("EleutherAI/pythia-410m-deduped" ) for checkpointing in [True, False]: a : List[str] = GPTNeoXForCausalLM.from_pretrained("EleutherAI/pythia-410m-deduped" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(lowerCAmelCase__ ) a : str = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCAmelCase__ ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 a : List[str] = "My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure" a : List[str] = model.generate(**lowerCAmelCase__ , do_sample=lowerCAmelCase__ , max_new_tokens=20 ) a : str = tokenizer.batch_decode(lowerCAmelCase__ )[0] self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ )
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" import os import string import sys a : Dict = 1 << 8 a : List[str] = { '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 27, '''up''': 65 + ARROW_KEY_FLAG, '''down''': 66 + ARROW_KEY_FLAG, '''right''': 67 + ARROW_KEY_FLAG, '''left''': 68 + ARROW_KEY_FLAG, '''mod_int''': 91, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 50, '''delete''': 51, '''pg_up''': 53, '''pg_down''': 54, } a : int = KEYMAP['''up'''] a : List[str] = KEYMAP['''left'''] if sys.platform == "win32": a : List[str] = [] a : Dict = { b'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, b'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(10): a : Any = ord(str(i)) def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' if os.name == "nt": import msvcrt a : Any = "mbcs" # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(_lowercase ) == 0: # Read the keystroke a : Dict = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): a : List[Any] = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: a : Optional[Any] = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP["mod_int"] ) ) WIN_CH_BUFFER.append(_lowercase ) if ord(_lowercase ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) a : List[str] = chr(KEYMAP["esc"] ) except KeyError: a : Any = cha[1] else: a : str = ch.decode(_lowercase ) else: a : int = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty a : Optional[int] = sys.stdin.fileno() a : List[Any] = termios.tcgetattr(_lowercase ) try: tty.setraw(_lowercase ) a : Optional[int] = sys.stdin.read(1 ) finally: termios.tcsetattr(_lowercase , termios.TCSADRAIN , _lowercase ) return ch def _SCREAMING_SNAKE_CASE ( ) ->Optional[Any]: '''simple docstring''' a : Tuple = get_raw_chars() if ord(_lowercase ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(_lowercase ) == KEYMAP["esc"]: a : Optional[int] = get_raw_chars() if ord(_lowercase ) == KEYMAP["mod_int"]: a : Optional[Any] = get_raw_chars() if ord(_lowercase ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(_lowercase ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(_lowercase ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : List[Any] = { '''tanreinama/GPTSAN-2.8B-spout_is_uniform''': ( '''https://huggingface.co/tanreinama/GPTSAN-2.8B-spout_is_uniform/resolve/main/config.json''' ), } class __UpperCamelCase ( a__ ): lowerCamelCase : List[Any] ="""gptsan-japanese""" lowerCamelCase : Any =[ """past_key_values""", ] lowerCamelCase : Optional[Any] ={ """hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , lowerCAmelCase__=3_6000 , lowerCAmelCase__=1280 , lowerCAmelCase__=1024 , lowerCAmelCase__=8192 , lowerCAmelCase__=4096 , lowerCAmelCase__=128 , lowerCAmelCase__=10 , lowerCAmelCase__=0 , lowerCAmelCase__=16 , lowerCAmelCase__=16 , lowerCAmelCase__=128 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=False , lowerCAmelCase__=0.0 , lowerCAmelCase__="float32" , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=0.002 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=3_5998 , lowerCAmelCase__=3_5995 , lowerCAmelCase__=3_5999 , **lowerCAmelCase__ , ) -> Optional[Any]: a : Tuple = vocab_size a : Any = max_position_embeddings a : Tuple = d_model a : Dict = d_ff a : Tuple = d_ext a : List[str] = d_spout a : Dict = num_switch_layers a : Any = num_ext_layers a : Union[str, Any] = num_switch_layers + num_ext_layers a : Dict = num_heads a : List[Any] = num_experts a : Dict = expert_capacity a : Optional[int] = dropout_rate a : int = layer_norm_epsilon a : Union[str, Any] = router_bias a : Any = router_jitter_noise a : List[Any] = router_dtype a : List[str] = router_ignore_padding_tokens a : str = output_hidden_states a : Any = output_attentions a : Optional[int] = initializer_factor a : Dict = output_router_logits a : Tuple = use_cache super().__init__( separator_token_id=lowerCAmelCase__ , pad_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging a : Union[str, Any] = logging.get_logger(__name__) class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Union[str, Any]: warnings.warn( "`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` " "instead." , lowerCAmelCase__ , ) super().__init__(args=lowerCAmelCase__ , **lowerCAmelCase__ )
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->Optional[int]: '''simple docstring''' if b == 0: return 1 if (b % 2) == 0: return actual_power(_lowercase , int(b / 2 ) ) * actual_power(_lowercase , int(b / 2 ) ) else: return a * actual_power(_lowercase , int(b / 2 ) ) * actual_power(_lowercase , int(b / 2 ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->float: '''simple docstring''' if b < 0: return 1 / actual_power(_lowercase , _lowercase ) return actual_power(_lowercase , _lowercase ) if __name__ == "__main__": print(power(-2, -3))
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a : Optional[int] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ['''GPTSw3Tokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys a : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" # Algorithm for the pigeonhole sorting def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Any: '''simple docstring''' a : Optional[int] = min(_lowercase ) # min() finds the minimum value a : str = max(_lowercase ) # max() finds the maximum value a : Dict = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size a : List[str] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(_lowercase , _lowercase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. a : Union[str, Any] = 0 for count in range(_lowercase ): while holes[count] > 0: holes[count] -= 1 a : List[Any] = count + min_val i += 1 def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Union[str, Any] = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(_lowercase ) print("Sorted order is:" , " ".join(_lowercase ) ) if __name__ == "__main__": main()
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCamelCase : lowerCamelCase : torch.Tensor # [batch_size x 3] lowerCamelCase : torch.Tensor # [batch_size x 3] lowerCamelCase : torch.Tensor # [batch_size x 3] lowerCamelCase : torch.Tensor # [batch_size x 3] lowerCamelCase : int lowerCamelCase : int lowerCamelCase : float lowerCamelCase : float lowerCamelCase : Tuple[int] def __a ( self ) -> Union[str, Any]: assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def __a ( self ) -> Any: return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def __a ( self ) -> str: return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def __a ( self ) -> torch.Tensor: a : int = torch.arange(self.height * self.width ) a : Optional[int] = torch.stack( [ pixel_indices % self.width, torch.div(lowerCAmelCase__ , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def __a ( self ) -> Optional[Any]: a, *a : Dict = self.shape a : Any = int(np.prod(lowerCAmelCase__ ) ) a : Optional[int] = self.get_image_coords() a : List[str] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) a : Dict = self.get_camera_rays(lowerCAmelCase__ ) a : Tuple = rays.view(lowerCAmelCase__ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def __a ( self , lowerCAmelCase__ ) -> torch.Tensor: a, *a, a : List[Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] a : Optional[int] = coords.view(lowerCAmelCase__ , -1 , 2 ) a : Dict = self.resolution() a : Optional[Any] = self.fov() a : Optional[int] = (flat.float() / (res - 1)) * 2 - 1 a : str = fracs * torch.tan(fov / 2 ) a : Tuple = fracs.view(lowerCAmelCase__ , -1 , 2 ) a : List[Any] = ( self.z.view(lowerCAmelCase__ , 1 , 3 ) + self.x.view(lowerCAmelCase__ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCAmelCase__ , 1 , 3 ) * fracs[:, :, 1:] ) a : Dict = directions / directions.norm(dim=-1 , keepdim=lowerCAmelCase__ ) a : Union[str, Any] = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCAmelCase__ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCAmelCase__ , *lowerCAmelCase__ , 2 , 3 ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> "DifferentiableProjectiveCamera": assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCAmelCase__ , height=lowerCAmelCase__ , x_fov=self.x_fov , y_fov=self.y_fov , ) def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->DifferentiableProjectiveCamera: '''simple docstring''' a : int = [] a : str = [] a : List[str] = [] a : Optional[Any] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): a : Optional[Any] = np.array([np.sin(_lowercase ), np.cos(_lowercase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) a : Union[str, Any] = -z * 4 a : Tuple = np.array([np.cos(_lowercase ), -np.sin(_lowercase ), 0.0] ) a : List[Any] = np.cross(_lowercase , _lowercase ) origins.append(_lowercase ) xs.append(_lowercase ) ys.append(_lowercase ) zs.append(_lowercase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , width=_lowercase , height=_lowercase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(_lowercase )) , )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" import os from collections.abc import Iterator def _SCREAMING_SNAKE_CASE ( _lowercase : str = "." ) ->Iterator[str]: '''simple docstring''' for dir_path, dir_names, filenames in os.walk(_lowercase ): a : str = [d for d in dir_names if d != "scripts" and d[0] not in "._"] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(_lowercase )[1] in (".py", ".ipynb"): yield os.path.join(_lowercase , _lowercase ).lstrip("./" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Tuple: '''simple docstring''' return F"""{i * ' '}*""" if i else "\n##" def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : str ) ->str: '''simple docstring''' a : str = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(_lowercase ) or old_parts[i] != new_part) and new_part: print(F"""{md_prefix(_lowercase )} {new_part.replace('_' , ' ' ).title()}""" ) return new_path def _SCREAMING_SNAKE_CASE ( _lowercase : str = "." ) ->None: '''simple docstring''' a : str = "" for filepath in sorted(good_file_paths(_lowercase ) ): a, a : List[Any] = os.path.split(_lowercase ) if filepath != old_path: a : Union[str, Any] = print_path(_lowercase , _lowercase ) a : List[str] = (filepath.count(os.sep ) + 1) if filepath else 0 a : str = F"""{filepath}/{filename}""".replace(" " , "%20" ) a : List[Any] = os.path.splitext(filename.replace("_" , " " ).title() )[0] print(F"""{md_prefix(_lowercase )} [{filename}]({url})""" ) if __name__ == "__main__": print_directory_md('''.''')
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" from maths.prime_check import is_prime def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->int: '''simple docstring''' if not isinstance(_lowercase , _lowercase ): a : Optional[int] = F"""Input value of [number={number}] must be an integer""" raise TypeError(_lowercase ) if is_prime(_lowercase ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" import math import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int = 1 , _lowercase : int = 1 , _lowercase : int = 1 ) ->qiskit.result.counts.Counts: '''simple docstring''' if ( isinstance(_lowercase , _lowercase ) or isinstance(_lowercase , _lowercase ) or isinstance(_lowercase , _lowercase ) ): raise TypeError("inputs must be integers." ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("inputs must be positive." ) if ( (math.floor(_lowercase ) != input_a) or (math.floor(_lowercase ) != input_a) or (math.floor(_lowercase ) != carry_in) ): raise ValueError("inputs must be exact integers." ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("inputs must be less or equal to 2." ) # build registers a : Optional[int] = qiskit.QuantumRegister(4 , "qr" ) a : Any = qiskit.ClassicalRegister(2 , "cr" ) # list the entries a : List[str] = [input_a, input_a, carry_in] a : List[str] = qiskit.QuantumCircuit(_lowercase , _lowercase ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(_lowercase ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(_lowercase ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(_lowercase ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , _lowercase ) # measure the last two qbits a : int = qiskit.Aer.get_backend("aer_simulator" ) a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total sum count for state is: {quantum_full_adder(1, 1, 1)}''')
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard @pytest.mark.parametrize( "compression_format, is_archive" , [ ("7z", True), ("bz2", False), ("gzip", False), ("lz4", False), ("tar", True), ("xz", False), ("zip", True), ("zstd", False), ] , ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Any , _lowercase : Union[str, Any] , _lowercase : List[str] , _lowercase : Union[str, Any] , _lowercase : Tuple , _lowercase : Dict , _lowercase : Optional[int] , _lowercase : List[str] , _lowercase : Union[str, Any] , _lowercase : str , _lowercase : int , ) ->str: '''simple docstring''' a : Optional[Any] = { "7z": (seven_zip_file, SevenZipExtractor), "bz2": (bza_file, BzipaExtractor), "gzip": (gz_file, GzipExtractor), "lz4": (lza_file, LzaExtractor), "tar": (tar_file, TarExtractor), "xz": (xz_file, XzExtractor), "zip": (zip_file, ZipExtractor), "zstd": (zstd_file, ZstdExtractor), } a, a : int = input_paths_and_base_extractors[compression_format] if input_path is None: a : List[str] = F"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_lowercase ) assert base_extractor.is_extractable(_lowercase ) a : Dict = tmp_path / ("extracted" if is_archive else "extracted.txt") base_extractor.extract(_lowercase , _lowercase ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name a : List[Any] = file_path.read_text(encoding="utf-8" ) else: a : Optional[Any] = output_path.read_text(encoding="utf-8" ) a : Optional[int] = text_file.read_text(encoding="utf-8" ) assert extracted_file_content == expected_file_content @pytest.mark.parametrize( "compression_format, is_archive" , [ ("7z", True), ("bz2", False), ("gzip", False), ("lz4", False), ("tar", True), ("xz", False), ("zip", True), ("zstd", False), ] , ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Optional[Any] , _lowercase : str , _lowercase : Dict , _lowercase : Union[str, Any] , _lowercase : List[Any] , _lowercase : str , _lowercase : Any , _lowercase : Optional[int] , _lowercase : Any , _lowercase : List[Any] , _lowercase : Optional[int] , ) ->Union[str, Any]: '''simple docstring''' a : str = { "7z": seven_zip_file, "bz2": bza_file, "gzip": gz_file, "lz4": lza_file, "tar": tar_file, "xz": xz_file, "zip": zip_file, "zstd": zstd_file, } a : List[Any] = input_paths[compression_format] if input_path is None: a : List[str] = F"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_lowercase ) a : str = Extractor.infer_extractor_format(_lowercase ) assert extractor_format is not None a : Dict = tmp_path / ("extracted" if is_archive else "extracted.txt") Extractor.extract(_lowercase , _lowercase , _lowercase ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name a : Optional[int] = file_path.read_text(encoding="utf-8" ) else: a : Optional[int] = output_path.read_text(encoding="utf-8" ) a : List[str] = text_file.read_text(encoding="utf-8" ) assert extracted_file_content == expected_file_content @pytest.fixture def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Optional[Any] ) ->Dict: '''simple docstring''' import tarfile a : int = tmp_path / "data_dot_dot" directory.mkdir() a : Optional[Any] = directory / "tar_file_with_dot_dot.tar" with tarfile.TarFile(_lowercase , "w" ) as f: f.add(_lowercase , arcname=os.path.join(".." , text_file.name ) ) return path @pytest.fixture def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->Optional[Any]: '''simple docstring''' import tarfile a : Dict = tmp_path / "data_sym_link" directory.mkdir() a : str = directory / "tar_file_with_sym_link.tar" os.symlink(".." , directory / "subdir" , target_is_directory=_lowercase ) with tarfile.TarFile(_lowercase , "w" ) as f: f.add(str(directory / "subdir" ) , arcname="subdir" ) # str required by os.readlink on Windows and Python < 3.8 return path @pytest.mark.parametrize( "insecure_tar_file, error_log" , [("tar_file_with_dot_dot", "illegal path"), ("tar_file_with_sym_link", "Symlink")] , ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : List[str] , _lowercase : int , _lowercase : int , _lowercase : str , _lowercase : int ) ->Optional[Any]: '''simple docstring''' a : Any = { "tar_file_with_dot_dot": tar_file_with_dot_dot, "tar_file_with_sym_link": tar_file_with_sym_link, } a : List[str] = insecure_tar_files[insecure_tar_file] a : str = tmp_path / "extracted" TarExtractor.extract(_lowercase , _lowercase ) assert caplog.text for record in caplog.records: assert record.levelname == "ERROR" assert error_log in record.msg def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->List[str]: '''simple docstring''' a : List[Any] = tmpdir / "not_a_zip_file" # From: https://github.com/python/cpython/pull/5053 a : Any = ( B"\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x01\x00\x00" B"\x00\x02\x08\x06\x00\x00\x00\x99\x81\xb6'\x00\x00\x00\x15I" B"DATx\x01\x01\n\x00\xf5\xff\x00PK\x05\x06\x00PK\x06\x06\x07" B"\xac\x01N\xc6|a\r\x00\x00\x00\x00IEND\xaeB`\x82" ) with not_a_zip_file.open("wb" ) as f: f.write(_lowercase ) assert zipfile.is_zipfile(str(_lowercase ) ) # is a false positive for `zipfile` assert not ZipExtractor.is_extractable(_lowercase ) # but we're right
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __UpperCamelCase ( a__ ): def __a ( self ) -> Optional[int]: a : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "tf_padding" ) ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "depth_multiplier" ) ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=3 , lowerCAmelCase__=32 , lowerCAmelCase__=0.25 , lowerCAmelCase__=8 , lowerCAmelCase__=8 , lowerCAmelCase__=6 , lowerCAmelCase__=32 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__="relu6" , lowerCAmelCase__=1280 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.02 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=10 , lowerCAmelCase__=None , ) -> str: a : Dict = parent a : int = batch_size a : List[Any] = num_channels a : List[Any] = image_size a : Dict = depth_multiplier a : str = depth_divisible_by a : str = min_depth a : Union[str, Any] = expand_ratio a : Tuple = tf_padding a : Dict = output_stride a : Optional[int] = first_layer_is_expansion a : Optional[Any] = finegrained_output a : List[str] = hidden_act a : List[str] = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) a : List[str] = classifier_dropout_prob a : Tuple = use_labels a : int = is_training a : Tuple = num_labels a : Union[str, Any] = initializer_range a : Dict = scope def __a ( self ) -> str: a : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a : str = None a : Union[str, Any] = None if self.use_labels: a : Dict = ids_tensor([self.batch_size] , self.num_labels ) a : Union[str, Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) a : List[str] = self.get_config() return config, pixel_values, labels, pixel_labels def __a ( self ) -> List[str]: return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : int = MobileNetVaModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Optional[int] = self.num_labels a : Tuple = MobileNetVaForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Dict = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Any = self.num_labels a : Tuple = MobileNetVaForSemanticSegmentation(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[Any] = model(lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) a : Union[str, Any] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __a ( self ) -> str: a : Union[str, Any] = self.prepare_config_and_inputs() a, a, a, a : List[str] = config_and_inputs a : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) lowerCamelCase : Optional[int] =( { """feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification, """image-segmentation""": MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) lowerCamelCase : Tuple =False lowerCamelCase : Dict =False lowerCamelCase : Optional[Any] =False lowerCamelCase : Optional[int] =False def __a ( self ) -> List[Any]: a : Union[str, Any] = MobileNetVaModelTester(self ) a : Any = MobileNetVaConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def __a ( self ) -> str: self.config_tester.run_common_tests() @unittest.skip(reason="MobileNetV2 does not use inputs_embeds" ) def __a ( self ) -> str: pass @unittest.skip(reason="MobileNetV2 does not support input and output embeddings" ) def __a ( self ) -> int: pass @unittest.skip(reason="MobileNetV2 does not output attentions" ) def __a ( self ) -> List[Any]: pass def __a ( self ) -> Optional[int]: a, a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Union[str, Any] = model_class(lowerCAmelCase__ ) a : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : Any = [*signature.parameters.keys()] a : Tuple = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def __a ( self ) -> Optional[int]: a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __a ( self ) -> Optional[int]: def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): a : Optional[Any] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = outputs.hidden_states a : int = 16 self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) a, a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Dict = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a : Union[str, Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[Any]: a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) def __a ( self ) -> Any: a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase__ ) @slow def __a ( self ) -> str: for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Tuple = MobileNetVaModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def __a ( self ) -> Optional[Any]: return ( MobileNetVaImageProcessor.from_pretrained("google/mobilenet_v2_1.0_224" ) if is_vision_available() else None ) @slow def __a ( self ) -> List[str]: a : Optional[Any] = MobileNetVaForImageClassification.from_pretrained("google/mobilenet_v2_1.0_224" ).to(lowerCAmelCase__ ) a : List[str] = self.default_image_processor a : List[Any] = prepare_img() a : Dict = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : Tuple = model(**lowerCAmelCase__ ) # verify the logits a : str = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) a : Tuple = torch.tensor([0.2_445, -1.1_993, 0.1_905] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def __a ( self ) -> int: a : int = MobileNetVaForSemanticSegmentation.from_pretrained("google/deeplabv3_mobilenet_v2_1.0_513" ) a : Optional[Any] = model.to(lowerCAmelCase__ ) a : List[Any] = MobileNetVaImageProcessor.from_pretrained("google/deeplabv3_mobilenet_v2_1.0_513" ) a : List[Any] = prepare_img() a : Optional[int] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : int = model(**lowerCAmelCase__ ) a : Optional[int] = outputs.logits # verify the logits a : Tuple = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , lowerCAmelCase__ ) a : Optional[int] = torch.tensor( [ [[17.5_790, 17.7_581, 18.3_355], [18.3_257, 18.4_230, 18.8_973], [18.6_169, 18.8_650, 19.2_187]], [[-2.1_595, -2.0_977, -2.3_741], [-2.4_226, -2.3_028, -2.6_835], [-2.7_819, -2.5_991, -2.7_706]], [[4.2_058, 4.8_317, 4.7_638], [4.4_136, 5.0_361, 4.9_383], [4.5_028, 4.9_644, 4.8_734]], ] , device=lowerCAmelCase__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) )
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__=3 , lowerCAmelCase__=3 , lowerCAmelCase__=("DownEncoderBlock2D",) , lowerCAmelCase__=(64,) , lowerCAmelCase__=2 , lowerCAmelCase__=32 , lowerCAmelCase__="silu" , lowerCAmelCase__=True , ) -> Dict: super().__init__() a : Tuple = layers_per_block a : List[str] = torch.nn.Convad( lowerCAmelCase__ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) a : str = None a : Any = nn.ModuleList([] ) # down a : Tuple = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase__ ): a : Dict = output_channel a : str = block_out_channels[i] a : int = i == len(lowerCAmelCase__ ) - 1 a : List[Any] = get_down_block( lowerCAmelCase__ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase__ , out_channels=lowerCAmelCase__ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase__ , resnet_groups=lowerCAmelCase__ , attention_head_dim=lowerCAmelCase__ , temb_channels=lowerCAmelCase__ , ) self.down_blocks.append(lowerCAmelCase__ ) # mid a : str = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase__ , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase__ , temb_channels=lowerCAmelCase__ , ) # out a : Tuple = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase__ , eps=1E-6 ) a : Dict = nn.SiLU() a : Union[str, Any] = 2 * out_channels if double_z else out_channels a : List[Any] = nn.Convad(block_out_channels[-1] , lowerCAmelCase__ , 3 , padding=1 ) a : List[str] = False def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: a : Optional[Any] = x a : Dict = self.conv_in(lowerCAmelCase__ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase__ ): def custom_forward(*lowerCAmelCase__ ): return module(*lowerCAmelCase__ ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: a : List[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase__ ) , lowerCAmelCase__ , use_reentrant=lowerCAmelCase__ ) # middle a : List[str] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase__ , use_reentrant=lowerCAmelCase__ ) else: for down_block in self.down_blocks: a : List[Any] = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase__ ) , lowerCAmelCase__ ) # middle a : Optional[int] = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase__ ) else: # down for down_block in self.down_blocks: a : Optional[int] = down_block(lowerCAmelCase__ ) # middle a : int = self.mid_block(lowerCAmelCase__ ) # post-process a : Tuple = self.conv_norm_out(lowerCAmelCase__ ) a : int = self.conv_act(lowerCAmelCase__ ) a : str = self.conv_out(lowerCAmelCase__ ) return sample class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__=3 , lowerCAmelCase__=3 , lowerCAmelCase__=("UpDecoderBlock2D",) , lowerCAmelCase__=(64,) , lowerCAmelCase__=2 , lowerCAmelCase__=32 , lowerCAmelCase__="silu" , lowerCAmelCase__="group" , ) -> Dict: super().__init__() a : List[str] = layers_per_block a : Optional[int] = nn.Convad( lowerCAmelCase__ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) a : int = None a : Union[str, Any] = nn.ModuleList([] ) a : Optional[int] = in_channels if norm_type == "spatial" else None # mid a : Any = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase__ , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase__ , temb_channels=lowerCAmelCase__ , ) # up a : List[str] = list(reversed(lowerCAmelCase__ ) ) a : List[Any] = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase__ ): a : str = output_channel a : Tuple = reversed_block_out_channels[i] a : Any = i == len(lowerCAmelCase__ ) - 1 a : str = get_up_block( lowerCAmelCase__ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase__ , out_channels=lowerCAmelCase__ , prev_output_channel=lowerCAmelCase__ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase__ , resnet_groups=lowerCAmelCase__ , attention_head_dim=lowerCAmelCase__ , temb_channels=lowerCAmelCase__ , resnet_time_scale_shift=lowerCAmelCase__ , ) self.up_blocks.append(lowerCAmelCase__ ) a : List[str] = output_channel # out if norm_type == "spatial": a : Optional[Any] = SpatialNorm(block_out_channels[0] , lowerCAmelCase__ ) else: a : Optional[Any] = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase__ , eps=1E-6 ) a : int = nn.SiLU() a : Union[str, Any] = nn.Convad(block_out_channels[0] , lowerCAmelCase__ , 3 , padding=1 ) a : int = False def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> Any: a : Any = z a : Union[str, Any] = self.conv_in(lowerCAmelCase__ ) a : Any = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase__ ): def custom_forward(*lowerCAmelCase__ ): return module(*lowerCAmelCase__ ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle a : Optional[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase__ , lowerCAmelCase__ , use_reentrant=lowerCAmelCase__ ) a : List[Any] = sample.to(lowerCAmelCase__ ) # up for up_block in self.up_blocks: a : str = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase__ ) , lowerCAmelCase__ , lowerCAmelCase__ , use_reentrant=lowerCAmelCase__ ) else: # middle a : Optional[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase__ , lowerCAmelCase__ ) a : int = sample.to(lowerCAmelCase__ ) # up for up_block in self.up_blocks: a : Dict = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase__ ) , lowerCAmelCase__ , lowerCAmelCase__ ) else: # middle a : List[Any] = self.mid_block(lowerCAmelCase__ , lowerCAmelCase__ ) a : int = sample.to(lowerCAmelCase__ ) # up for up_block in self.up_blocks: a : Optional[Any] = up_block(lowerCAmelCase__ , lowerCAmelCase__ ) # post-process if latent_embeds is None: a : str = self.conv_norm_out(lowerCAmelCase__ ) else: a : Optional[int] = self.conv_norm_out(lowerCAmelCase__ , lowerCAmelCase__ ) a : Any = self.conv_act(lowerCAmelCase__ ) a : Tuple = self.conv_out(lowerCAmelCase__ ) return sample class __UpperCamelCase ( nn.Module ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__="random" , lowerCAmelCase__=False , lowerCAmelCase__=True ) -> List[Any]: super().__init__() a : Union[str, Any] = n_e a : Optional[Any] = vq_embed_dim a : int = beta a : Any = legacy a : str = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) a : str = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) a : List[Any] = self.used.shape[0] a : Optional[int] = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": a : List[str] = self.re_embed a : List[str] = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: a : str = n_e a : Union[str, Any] = sane_index_shape def __a ( self , lowerCAmelCase__ ) -> str: a : List[Any] = inds.shape assert len(lowerCAmelCase__ ) > 1 a : Union[str, Any] = inds.reshape(ishape[0] , -1 ) a : Optional[Any] = self.used.to(lowerCAmelCase__ ) a : List[str] = (inds[:, :, None] == used[None, None, ...]).long() a : str = match.argmax(-1 ) a : int = match.sum(2 ) < 1 if self.unknown_index == "random": a : Optional[int] = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: a : Optional[Any] = self.unknown_index return new.reshape(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> int: a : str = inds.shape assert len(lowerCAmelCase__ ) > 1 a : Tuple = inds.reshape(ishape[0] , -1 ) a : Dict = self.used.to(lowerCAmelCase__ ) if self.re_embed > self.used.shape[0]: # extra token a : Optional[Any] = 0 # simply set to zero a : Optional[Any] = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase__ ) return back.reshape(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> int: # reshape z -> (batch, height, width, channel) and flatten a : Tuple = z.permute(0 , 2 , 3 , 1 ).contiguous() a : Optional[int] = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z a : Tuple = torch.argmin(torch.cdist(lowerCAmelCase__ , self.embedding.weight ) , dim=1 ) a : Dict = self.embedding(lowerCAmelCase__ ).view(z.shape ) a : Dict = None a : Tuple = None # compute loss for embedding if not self.legacy: a : Optional[int] = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: a : Union[str, Any] = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients a : Optional[int] = z + (z_q - z).detach() # reshape back to match original input shape a : Dict = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: a : Optional[Any] = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis a : Optional[Any] = self.remap_to_used(lowerCAmelCase__ ) a : List[Any] = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: a : List[Any] = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: # shape specifying (batch, height, width, channel) if self.remap is not None: a : List[str] = indices.reshape(shape[0] , -1 ) # add batch axis a : Optional[Any] = self.unmap_to_all(lowerCAmelCase__ ) a : Any = indices.reshape(-1 ) # flatten again # get quantized latent vectors a : str = self.embedding(lowerCAmelCase__ ) if shape is not None: a : List[str] = z_q.view(lowerCAmelCase__ ) # reshape back to match original input shape a : Optional[int] = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=False ) -> int: a : Any = parameters a, a : Tuple = torch.chunk(lowerCAmelCase__ , 2 , dim=1 ) a : Optional[Any] = torch.clamp(self.logvar , -30.0 , 20.0 ) a : int = deterministic a : Optional[Any] = torch.exp(0.5 * self.logvar ) a : Optional[int] = torch.exp(self.logvar ) if self.deterministic: a : int = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def __a ( self , lowerCAmelCase__ = None ) -> torch.FloatTensor: # make sure sample is on the same device as the parameters and has same dtype a : List[str] = randn_tensor( self.mean.shape , generator=lowerCAmelCase__ , device=self.parameters.device , dtype=self.parameters.dtype ) a : List[str] = self.mean + self.std * sample return x def __a ( self , lowerCAmelCase__=None ) -> Tuple: if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=[1, 2, 3] ) -> int: if self.deterministic: return torch.Tensor([0.0] ) a : List[str] = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase__ ) def __a ( self ) -> Any: return self.mean
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : str = "The quick brown fox jumps over the lazy dog" , ) ->bool: '''simple docstring''' a : str = set() # Replace all the whitespace in our sentence a : str = input_str.replace(" " , "" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(_lowercase ) == 26 def _SCREAMING_SNAKE_CASE ( _lowercase : str = "The quick brown fox jumps over the lazy dog" , ) ->bool: '''simple docstring''' a : Optional[Any] = [False] * 26 for char in input_str: if char.islower(): a : Tuple = True elif char.isupper(): a : Tuple = True return all(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : str = "The quick brown fox jumps over the lazy dog" , ) ->bool: '''simple docstring''' return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' from timeit import timeit a : Dict = "from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest" print(timeit("is_pangram()" , setup=_lowercase ) ) print(timeit("is_pangram_faster()" , setup=_lowercase ) ) print(timeit("is_pangram_fastest()" , setup=_lowercase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" from __future__ import annotations from functools import lru_cache from math import ceil a : int = 100 a : str = set(range(3, NUM_PRIMES, 2)) primes.add(2) a : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->set[int]: '''simple docstring''' if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} a : set[int] = set() a : int a : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def _SCREAMING_SNAKE_CASE ( _lowercase : int = 5000 ) ->int | None: '''simple docstring''' for number_to_partition in range(1 , _lowercase ): if len(partition(_lowercase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F'''{solution() = }''')
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml a : Dict = NewType('''DataClass''', Any) a : Tuple = NewType('''DataClassType''', Any) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->Optional[Any]: '''simple docstring''' if isinstance(_lowercase , _lowercase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def _SCREAMING_SNAKE_CASE ( _lowercase : list ) ->Callable[[str], Any]: '''simple docstring''' a : Dict = {str(_lowercase ): choice for choice in choices} return lambda _lowercase : str_to_choice.get(_lowercase , _lowercase ) def _SCREAMING_SNAKE_CASE ( *, _lowercase : Union[str, List[str]] = None , _lowercase : str = None , _lowercase : Any = dataclasses.MISSING , _lowercase : Callable[[], Any] = dataclasses.MISSING , _lowercase : dict = None , **_lowercase : Union[str, Any] , ) ->dataclasses.Field: '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls a : Tuple = {} if aliases is not None: a : Union[str, Any] = aliases if help is not None: a : Optional[Any] = help return dataclasses.field(metadata=_lowercase , default=_lowercase , default_factory=_lowercase , **_lowercase ) class __UpperCamelCase ( a__ ): lowerCamelCase : Iterable[DataClassType] def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: # To make the default appear when using --help if "formatter_class" not in kwargs: a : Any = ArgumentDefaultsHelpFormatter super().__init__(**lowerCAmelCase__ ) if dataclasses.is_dataclass(lowerCAmelCase__ ): a : Dict = [dataclass_types] a : List[str] = list(lowerCAmelCase__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : int = f"""--{field.name}""" a : List[Any] = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , lowerCAmelCase__ ): raise RuntimeError( "Unresolved type detected, which should have been done with the help of " "`typing.get_type_hints` method by default" ) a : Dict = kwargs.pop("aliases" , [] ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Optional[Any] = [aliases] a : Union[str, Any] = getattr(field.type , "__origin__" , field.type ) if origin_type is Union or (hasattr(lowerCAmelCase__ , "UnionType" ) and isinstance(lowerCAmelCase__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowerCAmelCase__ ) not in field.type.__args__ ): raise ValueError( "Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because" " the argument parser only supports one type per argument." f""" Problem encountered in field '{field.name}'.""" ) if type(lowerCAmelCase__ ) not in field.type.__args__: # filter `str` in Union a : List[Any] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] a : str = getattr(field.type , "__origin__" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) a : str = ( field.type.__args__[0] if isinstance(lowerCAmelCase__ , field.type.__args__[1] ) else field.type.__args__[1] ) a : Union[str, Any] = getattr(field.type , "__origin__" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) a : List[Any] = {} if origin_type is Literal or (isinstance(field.type , lowerCAmelCase__ ) and issubclass(field.type , lowerCAmelCase__ )): if origin_type is Literal: a : Union[str, Any] = field.type.__args__ else: a : Optional[int] = [x.value for x in field.type] a : List[Any] = make_choice_type_function(kwargs["choices"] ) if field.default is not dataclasses.MISSING: a : Any = field.default else: a : Optional[int] = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument a : List[Any] = copy(lowerCAmelCase__ ) # Hack because type=bool in argparse does not behave as we want. a : Any = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. a : int = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way a : List[Any] = default # This tells argparse we accept 0 or 1 value after --field_name a : str = "?" # This is the value that will get picked if we do --field_name (without value) a : Optional[Any] = True elif isclass(lowerCAmelCase__ ) and issubclass(lowerCAmelCase__ , lowerCAmelCase__ ): a : int = field.type.__args__[0] a : str = "+" if field.default_factory is not dataclasses.MISSING: a : Tuple = field.default_factory() elif field.default is dataclasses.MISSING: a : List[str] = True else: a : str = field.type if field.default is not dataclasses.MISSING: a : Union[str, Any] = field.default elif field.default_factory is not dataclasses.MISSING: a : Optional[int] = field.default_factory() else: a : Tuple = True parser.add_argument(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): a : Dict = False parser.add_argument(f"""--no_{field.name}""" , action="store_false" , dest=field.name , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> List[str]: if hasattr(lowerCAmelCase__ , "_argument_group_name" ): a : Any = self.add_argument_group(dtype._argument_group_name ) else: a : Optional[Any] = self try: a : Dict[str, type] = get_type_hints(lowerCAmelCase__ ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ "removing line of `from __future__ import annotations` which opts in Postponed " "Evaluation of Annotations (PEP 563)" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(lowerCAmelCase__ ): a : List[Any] = ".".join(map(lowerCAmelCase__ , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ "line of `from __future__ import annotations` which opts in union types as " "`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To " "support Python versions that lower than 3.10, you need to use " "`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of " "`X | None`." ) from ex raise for field in dataclasses.fields(lowerCAmelCase__ ): if not field.init: continue a : Any = type_hints[field.name] self._parse_dataclass_field(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): a : Dict = [] if args_filename: args_files.append(Path(lowerCAmelCase__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(".args" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values a : int = ArgumentParser() args_file_parser.add_argument(lowerCAmelCase__ , type=lowerCAmelCase__ , action="append" ) # Use only remaining args for further parsing (remove the args_file_flag) a, a : Dict = args_file_parser.parse_known_args(args=lowerCAmelCase__ ) a : Optional[int] = vars(lowerCAmelCase__ ).get(args_file_flag.lstrip("-" ) , lowerCAmelCase__ ) if cmd_args_file_paths: args_files.extend([Path(lowerCAmelCase__ ) for p in cmd_args_file_paths] ) a : Union[str, Any] = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last a : Any = file_args + args if args is not None else file_args + sys.argv[1:] a, a : List[str] = self.parse_known_args(args=lowerCAmelCase__ ) a : Dict = [] for dtype in self.dataclass_types: a : Any = {f.name for f in dataclasses.fields(lowerCAmelCase__ ) if f.init} a : Any = {k: v for k, v in vars(lowerCAmelCase__ ).items() if k in keys} for k in keys: delattr(lowerCAmelCase__ , lowerCAmelCase__ ) a : Any = dtype(**lowerCAmelCase__ ) outputs.append(lowerCAmelCase__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowerCAmelCase__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> Tuple[DataClass, ...]: a : Union[str, Any] = set(args.keys() ) a : List[str] = [] for dtype in self.dataclass_types: a : int = {f.name for f in dataclasses.fields(lowerCAmelCase__ ) if f.init} a : Tuple = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) a : List[str] = dtype(**lowerCAmelCase__ ) outputs.append(lowerCAmelCase__ ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(lowerCAmelCase__ )}""" ) return tuple(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> Tuple[DataClass, ...]: with open(Path(lowerCAmelCase__ ) , encoding="utf-8" ) as open_json_file: a : List[Any] = json.loads(open_json_file.read() ) a : Optional[Any] = self.parse_dict(lowerCAmelCase__ , allow_extra_keys=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> Tuple[DataClass, ...]: a : int = self.parse_dict(yaml.safe_load(Path(lowerCAmelCase__ ).read_text() ) , allow_extra_keys=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ )
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Dict = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Tuple ) ->Any: '''simple docstring''' a : List[str] = b.T a : Tuple = np.sum(np.square(_lowercase ) , axis=1 ) a : Optional[Any] = np.sum(np.square(_lowercase ) , axis=0 ) a : str = np.matmul(_lowercase , _lowercase ) a : List[Any] = aa[:, None] - 2 * ab + ba[None, :] return d def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Union[str, Any] ) ->int: '''simple docstring''' a : List[str] = x.reshape(-1 , 3 ) a : int = squared_euclidean_distance(_lowercase , _lowercase ) return np.argmin(_lowercase , axis=1 ) class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =["""pixel_values"""] def __init__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = PILImageResampling.BILINEAR , lowerCAmelCase__ = True , lowerCAmelCase__ = True , **lowerCAmelCase__ , ) -> None: super().__init__(**lowerCAmelCase__ ) a : Any = size if size is not None else {"height": 256, "width": 256} a : Any = get_size_dict(lowerCAmelCase__ ) a : int = np.array(lowerCAmelCase__ ) if clusters is not None else None a : int = do_resize a : Any = size a : int = resample a : int = do_normalize a : List[str] = do_color_quantize def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = PILImageResampling.BILINEAR , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: a : Dict = get_size_dict(lowerCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f"""Size dictionary must contain both height and width keys. Got {size.keys()}""" ) return resize( lowerCAmelCase__ , size=(size["height"], size["width"]) , resample=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , ) -> np.ndarray: a : Tuple = rescale(image=lowerCAmelCase__ , scale=1 / 127.5 , data_format=lowerCAmelCase__ ) a : Union[str, Any] = image - 1 return image def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = ChannelDimension.FIRST , **lowerCAmelCase__ , ) -> PIL.Image.Image: a : int = do_resize if do_resize is not None else self.do_resize a : List[Any] = size if size is not None else self.size a : Tuple = get_size_dict(lowerCAmelCase__ ) a : Dict = resample if resample is not None else self.resample a : Tuple = do_normalize if do_normalize is not None else self.do_normalize a : str = do_color_quantize if do_color_quantize is not None else self.do_color_quantize a : Optional[Any] = clusters if clusters is not None else self.clusters a : Optional[int] = np.array(lowerCAmelCase__ ) a : int = make_list_of_images(lowerCAmelCase__ ) if not valid_images(lowerCAmelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_color_quantize and clusters is None: raise ValueError("Clusters must be specified if do_color_quantize is True." ) # All transformations expect numpy arrays. a : Optional[Any] = [to_numpy_array(lowerCAmelCase__ ) for image in images] if do_resize: a : Tuple = [self.resize(image=lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ ) for image in images] if do_normalize: a : Optional[int] = [self.normalize(image=lowerCAmelCase__ ) for image in images] if do_color_quantize: a : str = [to_channel_dimension_format(lowerCAmelCase__ , ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) a : Optional[Any] = np.array(lowerCAmelCase__ ) a : Optional[int] = color_quantize(lowerCAmelCase__ , lowerCAmelCase__ ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) a : int = images.shape[0] a : int = images.reshape(lowerCAmelCase__ , -1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. a : Optional[Any] = list(lowerCAmelCase__ ) else: a : List[str] = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__ ) for image in images] a : Optional[int] = {"input_ids": images} return BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__ )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import math def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = input("Enter message: " ) a : List[Any] = int(input(F"""Enter key [2-{len(_lowercase ) - 1}]: """ ) ) a : Dict = input("Encryption/Decryption [e/d]: " ) if mode.lower().startswith("e" ): a : Dict = encrypt_message(_lowercase , _lowercase ) elif mode.lower().startswith("d" ): a : Optional[Any] = decrypt_message(_lowercase , _lowercase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F"""Output:\n{text + '|'}""" ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : str ) ->str: '''simple docstring''' a : str = [""] * key for col in range(_lowercase ): a : Optional[Any] = col while pointer < len(_lowercase ): cipher_text[col] += message[pointer] pointer += key return "".join(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : str ) ->str: '''simple docstring''' a : int = math.ceil(len(_lowercase ) / key ) a : Union[str, Any] = key a : List[Any] = (num_cols * num_rows) - len(_lowercase ) a : List[str] = [""] * num_cols a : int = 0 a : Any = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): a : List[Any] = 0 row += 1 return "".join(_lowercase ) if __name__ == "__main__": import doctest doctest.testmod() main()
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values a : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--user''', type=str, default='''ubuntu''') parser.add_argument('''--host''', type=str, default='''localhost''') parser.add_argument('''--key_path''', type=str, default=None) parser.add_argument('''--instance''', type=str, default='''V100:1''') parser.add_argument('''--provider''', type=str, default='''cheapest''') parser.add_argument('''--use_spot''', type=bool, default=False) parser.add_argument('''--example''', type=str, default='''pytorch/text-generation/run_generation.py''') a , a : Dict = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('''Cannot specify both BYO and on-demand cluster args''') a : Any = rh.cluster( name='''rh-cluster''', ips=[args.host], ssh_creds={'''ssh_user''': args.user, '''ssh_private_key''': args.key_path} ) else: a : List[str] = rh.cluster( name='''rh-cluster''', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) a : Optional[Any] = args.example.rsplit('''/''', 1)[0] # Set up remote environment cluster.install_packages(['''pip:./''']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F'''pip install -r transformers/examples/{example_dir}/requirements.txt''']) cluster.run(['''pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117''']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F'''python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}''']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" from __future__ import annotations import time a : Optional[Any] = list[tuple[int, int]] a : str = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] a : Dict = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : int = pos_x a : str = pos_y a : Tuple = (pos_y, pos_x) a : Optional[int] = goal_x a : Optional[Any] = goal_y a : List[Any] = parent class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: a : List[str] = Node(start[1] , start[0] , goal[1] , goal[0] , lowerCAmelCase__ ) a : Union[str, Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , lowerCAmelCase__ ) a : Any = [self.start] a : Any = False def __a ( self ) -> Path | None: while self.node_queue: a : Optional[int] = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: a : str = True return self.retrace_path(lowerCAmelCase__ ) a : Union[str, Any] = self.get_successors(lowerCAmelCase__ ) for node in successors: self.node_queue.append(lowerCAmelCase__ ) if not self.reached: return [self.start.pos] return None def __a ( self , lowerCAmelCase__ ) -> list[Node]: a : int = [] for action in delta: a : str = parent.pos_x + action[1] a : Optional[Any] = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCAmelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(lowerCAmelCase__ , lowerCAmelCase__ , self.target.pos_y , self.target.pos_x , lowerCAmelCase__ ) ) return successors def __a ( self , lowerCAmelCase__ ) -> Path: a : int = node a : Optional[int] = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) a : str = current_node.parent path.reverse() return path class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : Tuple = BreadthFirstSearch(lowerCAmelCase__ , lowerCAmelCase__ ) a : str = BreadthFirstSearch(lowerCAmelCase__ , lowerCAmelCase__ ) a : int = False def __a ( self ) -> Path | None: while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: a : Tuple = self.fwd_bfs.node_queue.pop(0 ) a : List[Any] = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: a : Dict = True return self.retrace_bidirectional_path( lowerCAmelCase__ , lowerCAmelCase__ ) a : Any = current_bwd_node a : List[str] = current_fwd_node a : Optional[Any] = { self.fwd_bfs: self.fwd_bfs.get_successors(lowerCAmelCase__ ), self.bwd_bfs: self.bwd_bfs.get_successors(lowerCAmelCase__ ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(lowerCAmelCase__ ) if not self.reached: return [self.fwd_bfs.start.pos] return None def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Path: a : str = self.fwd_bfs.retrace_path(lowerCAmelCase__ ) a : List[Any] = self.bwd_bfs.retrace_path(lowerCAmelCase__ ) bwd_path.pop() bwd_path.reverse() a : int = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() a : Optional[Any] = (0, 0) a : int = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) a : Optional[int] = time.time() a : Optional[Any] = BreadthFirstSearch(init, goal) a : Tuple = bfs.search() a : int = time.time() - start_bfs_time print('''Unidirectional BFS computation time : ''', bfs_time) a : Optional[Any] = time.time() a : Union[str, Any] = BidirectionalBreadthFirstSearch(init, goal) a : str = bd_bfs.search() a : List[str] = time.time() - start_bd_bfs_time print('''Bidirectional BFS computation time : ''', bd_bfs_time)
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : list[list] ) ->list[list]: '''simple docstring''' a : Union[str, Any] = current_set.copy() for row_index, row in enumerate(_lowercase ): a : Optional[int] = row[0] for column_index, column in enumerate(_lowercase ): if magnitude == 0: a : Optional[Any] = column continue a : str = column / magnitude # Subtract to cancel term a : Union[str, Any] = current_set[0] a : Optional[Any] = [first_row] a : List[Any] = current_set[1::] for row in current_set: a : Any = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(_lowercase ) continue for column_index in range(len(_lowercase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(_lowercase ) # Create next recursion iteration set if len(final_set[0] ) != 3: a : Tuple = final_set[0] a : Any = [] a : str = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) a : List[str] = simplify(_lowercase ) for i in range(len(_lowercase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , _lowercase ) a : Union[str, Any] = resultant return final_set def _SCREAMING_SNAKE_CASE ( _lowercase : list[list] ) ->list: '''simple docstring''' if len(_lowercase ) == 0: raise IndexError("solve_simultaneous() requires n lists of length n+1" ) a : Optional[int] = len(_lowercase ) + 1 if any(len(_lowercase ) != _length for item in equations ): raise IndexError("solve_simultaneous() requires n lists of length n+1" ) for row in equations: if any(not isinstance(_lowercase , (int, float) ) for column in row ): raise ValueError("solve_simultaneous() requires lists of integers" ) if len(_lowercase ) == 1: return [equations[0][-1] / equations[0][0]] a : Tuple = equations.copy() if any(0 in row for row in data_set ): a : List[str] = data_set.copy() a : List[str] = [] for row_index, row in enumerate(_lowercase ): if 0 not in row: a : List[Any] = data_set.pop(_lowercase ) break if not full_row: raise ValueError("solve_simultaneous() requires at least 1 full equation" ) data_set.insert(0 , _lowercase ) a : int = data_set.copy() a : List[str] = simplify(_lowercase ) a : Tuple = simplified[::-1] a : list = [] for row in simplified: a : str = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue a : Optional[Any] = row.copy()[: len(_lowercase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(_lowercase ) == 0: solutions.append(0 ) continue a : Dict = temp_row[1::] a : Union[str, Any] = temp_row[::-1] for column_index, column in enumerate(_lowercase ): current_solution -= column * solutions[column_index] solutions.append(_lowercase ) a : List[Any] = [] for item in solutions: final.append(float(round(_lowercase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() a : Tuple = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , ) -> List[str]: a : Tuple = parent a : str = 13 a : Any = 7 a : List[str] = True a : Optional[Any] = True a : Any = False a : Any = True a : Any = 99 a : Any = 32 a : Any = 2 a : List[Any] = 4 a : Union[str, Any] = 37 a : Dict = "gelu" a : List[Any] = 0.1 a : int = 0.1 a : Optional[int] = 512 a : str = 16 a : Dict = 2 a : Optional[int] = 0.02 a : List[Any] = 3 a : Union[str, Any] = 4 a : Optional[int] = None def __a ( self ) -> Dict: a : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : Any = None if self.use_input_mask: a : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) a : List[str] = None a : List[Any] = None a : Dict = None if self.use_labels: a : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) a : Union[str, Any] = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : List[Any] = TFDistilBertModel(config=lowerCAmelCase__ ) a : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask} a : Tuple = model(lowerCAmelCase__ ) a : List[Any] = [input_ids, input_mask] a : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: a : List[Any] = TFDistilBertForMaskedLM(config=lowerCAmelCase__ ) a : List[str] = {"input_ids": input_ids, "attention_mask": input_mask} a : Union[str, Any] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Any = TFDistilBertForQuestionAnswering(config=lowerCAmelCase__ ) a : int = { "input_ids": input_ids, "attention_mask": input_mask, } a : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: a : Dict = self.num_labels a : Optional[int] = TFDistilBertForSequenceClassification(lowerCAmelCase__ ) a : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask} a : Any = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Dict = self.num_choices a : Union[str, Any] = TFDistilBertForMultipleChoice(lowerCAmelCase__ ) a : Union[str, Any] = tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) a : List[str] = tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) a : List[Any] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } a : List[str] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: a : Dict = self.num_labels a : Tuple = TFDistilBertForTokenClassification(lowerCAmelCase__ ) a : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} a : Optional[int] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __a ( self ) -> Optional[int]: a : List[Any] = self.prepare_config_and_inputs() ((a), (a), (a), (a), (a), (a)) : Union[str, Any] = config_and_inputs a : List[str] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Optional[Any] =( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) lowerCamelCase : Any =( { """feature-extraction""": TFDistilBertModel, """fill-mask""": TFDistilBertForMaskedLM, """question-answering""": TFDistilBertForQuestionAnswering, """text-classification""": TFDistilBertForSequenceClassification, """token-classification""": TFDistilBertForTokenClassification, """zero-shot""": TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) lowerCamelCase : List[str] =False lowerCamelCase : Optional[int] =False def __a ( self ) -> str: a : Optional[Any] = TFDistilBertModelTester(self ) a : List[str] = ConfigTester(self , config_class=lowerCAmelCase__ , dim=37 ) def __a ( self ) -> Any: self.config_tester.run_common_tests() def __a ( self ) -> Union[str, Any]: a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*lowerCAmelCase__ ) def __a ( self ) -> Dict: a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*lowerCAmelCase__ ) def __a ( self ) -> Tuple: a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*lowerCAmelCase__ ) def __a ( self ) -> Dict: a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*lowerCAmelCase__ ) @slow def __a ( self ) -> Union[str, Any]: for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): a : Tuple = TFDistilBertModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @require_tf class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Tuple: a : List[Any] = TFDistilBertModel.from_pretrained("distilbert-base-uncased" ) a : Dict = tf.constant([[0, 1, 2, 3, 4, 5]] ) a : int = model(lowerCAmelCase__ )[0] a : int = [1, 6, 768] self.assertEqual(output.shape , lowerCAmelCase__ ) a : List[Any] = tf.constant( [ [ [0.19_261_885, -0.13_732_955, 0.4_119_799], [0.22_150_156, -0.07_422_661, 0.39_037_204], [0.22_756_018, -0.0_896_414, 0.3_701_467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCAmelCase__ , atol=1E-4 )
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" import sys a : Any = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->int: '''simple docstring''' a : Tuple = 1 for digit in s: product *= int(_lowercase ) return product def _SCREAMING_SNAKE_CASE ( _lowercase : str = N ) ->int: '''simple docstring''' a : Dict = -sys.maxsize - 1 a : Dict = n[:13] a : int = 13 while cur_index < len(_lowercase ) - 13: if int(n[cur_index] ) >= int(substr[0] ): a : Any = substr[1:] + n[cur_index] cur_index += 1 else: a : List[str] = max(_lowercase , str_eval(_lowercase ) ) a : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(F'''{solution() = }''')
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : int = logging.get_logger(__name__) a : Dict = { '''roberta-base''': '''https://huggingface.co/roberta-base/resolve/main/config.json''', '''roberta-large''': '''https://huggingface.co/roberta-large/resolve/main/config.json''', '''roberta-large-mnli''': '''https://huggingface.co/roberta-large-mnli/resolve/main/config.json''', '''distilroberta-base''': '''https://huggingface.co/distilroberta-base/resolve/main/config.json''', '''roberta-base-openai-detector''': '''https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json''', '''roberta-large-openai-detector''': '''https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json''', } class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] ="""roberta""" def __init__( self , lowerCAmelCase__=5_0265 , lowerCAmelCase__=768 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=3072 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__="absolute" , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> Optional[Any]: super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a : str = vocab_size a : Optional[Any] = hidden_size a : int = num_hidden_layers a : List[str] = num_attention_heads a : Union[str, Any] = hidden_act a : Optional[int] = intermediate_size a : Optional[Any] = hidden_dropout_prob a : Any = attention_probs_dropout_prob a : List[Any] = max_position_embeddings a : List[str] = type_vocab_size a : Optional[Any] = initializer_range a : List[str] = layer_norm_eps a : List[str] = position_embedding_type a : List[str] = use_cache a : str = classifier_dropout class __UpperCamelCase ( a__ ): @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": a : Union[str, Any] = {0: "batch", 1: "choice", 2: "sequence"} else: a : List[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig a : Optional[int] = logging.get_logger(__name__) # General docstring a : Tuple = '''RegNetConfig''' # Base docstring a : Any = '''facebook/regnet-y-040''' a : int = [1, 1088, 7, 7] # Image classification docstring a : Any = '''facebook/regnet-y-040''' a : Union[str, Any] = '''tabby, tabby cat''' a : Optional[Any] = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 1 , lowerCAmelCase__ = "relu" , **lowerCAmelCase__ , ) -> List[Any]: super().__init__(**lowerCAmelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb a : Optional[int] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) a : Optional[Any] = tf.keras.layers.ConvaD( filters=lowerCAmelCase__ , kernel_size=lowerCAmelCase__ , strides=lowerCAmelCase__ , padding="VALID" , groups=lowerCAmelCase__ , use_bias=lowerCAmelCase__ , name="convolution" , ) a : str = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) a : Dict = ACTaFN[activation] if activation is not None else tf.identity def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : Any = self.convolution(self.padding(lowerCAmelCase__ ) ) a : Optional[int] = self.normalization(lowerCAmelCase__ ) a : List[str] = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: super().__init__(**lowerCAmelCase__ ) a : Tuple = config.num_channels a : Optional[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : Union[str, Any] = shape_list(lowerCAmelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) a : List[str] = tf.transpose(lowerCAmelCase__ , perm=(0, 2, 3, 1) ) a : Optional[Any] = self.embedder(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = 2 , **lowerCAmelCase__ ) -> List[str]: super().__init__(**lowerCAmelCase__ ) a : Union[str, Any] = tf.keras.layers.ConvaD( filters=lowerCAmelCase__ , kernel_size=1 , strides=lowerCAmelCase__ , use_bias=lowerCAmelCase__ , name="convolution" ) a : Dict = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> tf.Tensor: return self.normalization(self.convolution(lowerCAmelCase__ ) , training=lowerCAmelCase__ ) class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[int]: super().__init__(**lowerCAmelCase__ ) a : str = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCAmelCase__ , name="pooler" ) a : Optional[int] = [ tf.keras.layers.ConvaD(filters=lowerCAmelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCAmelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] a : Dict = self.pooler(lowerCAmelCase__ ) for layer_module in self.attention: a : Optional[Any] = layer_module(lowerCAmelCase__ ) a : int = hidden_state * pooled return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 , **lowerCAmelCase__ ) -> int: super().__init__(**lowerCAmelCase__ ) a : Optional[int] = in_channels != out_channels or stride != 1 a : Dict = max(1 , out_channels // config.groups_width ) a : Optional[Any] = ( TFRegNetShortCut(lowerCAmelCase__ , stride=lowerCAmelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. a : Dict = [ TFRegNetConvLayer(lowerCAmelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCAmelCase__ , stride=lowerCAmelCase__ , groups=lowerCAmelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCAmelCase__ , kernel_size=1 , activation=lowerCAmelCase__ , name="layer.2" ), ] a : List[str] = ACTaFN[config.hidden_act] def __a ( self , lowerCAmelCase__ ) -> Optional[Any]: a : Any = hidden_state for layer_module in self.layers: a : Tuple = layer_module(lowerCAmelCase__ ) a : Dict = self.shortcut(lowerCAmelCase__ ) hidden_state += residual a : int = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 , **lowerCAmelCase__ ) -> List[Any]: super().__init__(**lowerCAmelCase__ ) a : List[str] = in_channels != out_channels or stride != 1 a : Any = max(1 , out_channels // config.groups_width ) a : str = ( TFRegNetShortCut(lowerCAmelCase__ , stride=lowerCAmelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) a : Optional[int] = [ TFRegNetConvLayer(lowerCAmelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCAmelCase__ , stride=lowerCAmelCase__ , groups=lowerCAmelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCAmelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCAmelCase__ , kernel_size=1 , activation=lowerCAmelCase__ , name="layer.3" ), ] a : Union[str, Any] = ACTaFN[config.hidden_act] def __a ( self , lowerCAmelCase__ ) -> Any: a : Union[str, Any] = hidden_state for layer_module in self.layers: a : str = layer_module(lowerCAmelCase__ ) a : Union[str, Any] = self.shortcut(lowerCAmelCase__ ) hidden_state += residual a : Optional[int] = self.activation(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 2 , **lowerCAmelCase__ ) -> List[Any]: super().__init__(**lowerCAmelCase__ ) a : Tuple = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer a : Optional[Any] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , stride=lowerCAmelCase__ , name="layers.0" ), *[layer(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , name=f"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def __a ( self , lowerCAmelCase__ ) -> int: for layer_module in self.layers: a : Dict = layer_module(lowerCAmelCase__ ) return hidden_state class __UpperCamelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: super().__init__(**lowerCAmelCase__ ) a : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCAmelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) a : Optional[Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCAmelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , depth=lowerCAmelCase__ , name=f"""stages.{i+1}""" ) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = False , lowerCAmelCase__ = True ) -> TFBaseModelOutputWithNoAttention: a : int = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: a : Optional[int] = hidden_states + (hidden_state,) a : Any = stage_module(lowerCAmelCase__ ) if output_hidden_states: a : Union[str, Any] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCAmelCase__ , hidden_states=lowerCAmelCase__ ) @keras_serializable class __UpperCamelCase ( tf.keras.layers.Layer ): lowerCamelCase : Tuple =RegNetConfig def __init__( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: super().__init__(**lowerCAmelCase__ ) a : int = config a : str = TFRegNetEmbeddings(lowerCAmelCase__ , name="embedder" ) a : Union[str, Any] = TFRegNetEncoder(lowerCAmelCase__ , name="encoder" ) a : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCAmelCase__ , name="pooler" ) @unpack_inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: a : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : int = return_dict if return_dict is not None else self.config.use_return_dict a : Any = self.embedder(lowerCAmelCase__ , training=lowerCAmelCase__ ) a : Tuple = self.encoder( lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , training=lowerCAmelCase__ ) a : int = encoder_outputs[0] a : List[Any] = self.pooler(lowerCAmelCase__ ) # Change to NCHW output format have uniformity in the modules a : Tuple = tf.transpose(lowerCAmelCase__ , perm=(0, 3, 1, 2) ) a : Optional[Any] = tf.transpose(lowerCAmelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: a : Tuple = tuple([tf.transpose(lowerCAmelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCAmelCase__ , pooler_output=lowerCAmelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class __UpperCamelCase ( a__ ): lowerCamelCase : List[str] =RegNetConfig lowerCamelCase : Any ="""regnet""" lowerCamelCase : int ="""pixel_values""" @property def __a ( self ) -> str: return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} a : Optional[Any] = R''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' a : Dict = R''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare RegNet model outputting raw features without any specific head on top.""" , a__ , ) class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: super().__init__(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) a : Union[str, Any] = TFRegNetMainLayer(lowerCAmelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: a : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict a : List[str] = self.regnet( pixel_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , training=lowerCAmelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( """ RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , a__ , ) class __UpperCamelCase ( a__ , a__ ): def __init__( self , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> str: super().__init__(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) a : Optional[Any] = config.num_labels a : Any = TFRegNetMainLayer(lowerCAmelCase__ , name="regnet" ) # classification head a : List[str] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCAmelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __a ( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: a : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : List[str] = return_dict if return_dict is not None else self.config.use_return_dict a : Union[str, Any] = self.regnet( lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , training=lowerCAmelCase__ ) a : Dict = outputs.pooler_output if return_dict else outputs[1] a : List[Any] = self.classifier[0](lowerCAmelCase__ ) a : Dict = self.classifier[1](lowerCAmelCase__ ) a : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCAmelCase__ , logits=lowerCAmelCase__ ) if not return_dict: a : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCAmelCase__ , logits=lowerCAmelCase__ , hidden_states=outputs.hidden_states )
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=14 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ) -> Any: a : int = parent a : List[str] = batch_size a : int = seq_length a : Optional[Any] = is_training a : List[Any] = use_token_type_ids a : int = use_input_mask a : Dict = use_labels a : Tuple = use_mc_token_ids a : Any = vocab_size a : Optional[Any] = hidden_size a : Optional[int] = num_hidden_layers a : Optional[int] = num_attention_heads a : List[Any] = intermediate_size a : Any = hidden_act a : Optional[int] = hidden_dropout_prob a : List[str] = attention_probs_dropout_prob a : Any = max_position_embeddings a : Optional[int] = type_vocab_size a : int = type_sequence_label_size a : Dict = initializer_range a : Optional[int] = num_labels a : Dict = num_choices a : List[str] = scope a : str = self.vocab_size - 1 def __a ( self ) -> Union[str, Any]: a : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : int = None if self.use_input_mask: a : int = random_attention_mask([self.batch_size, self.seq_length] ) a : Any = None if self.use_token_type_ids: a : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a : Optional[Any] = None if self.use_mc_token_ids: a : Tuple = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) a : Any = None a : Union[str, Any] = None a : Dict = None if self.use_labels: a : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) a : int = self.get_config() a : Any = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def __a ( self ) -> Optional[Any]: return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Dict: a : Union[str, Any] = CTRLModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , head_mask=lowerCAmelCase__ ) model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ ) a : str = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Any: a : Any = CTRLLMHeadModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : int = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self ) -> Dict: a : Union[str, Any] = self.prepare_config_and_inputs() ( ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ) : Optional[Any] = config_and_inputs a : Optional[Any] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "head_mask": head_mask} return config, inputs_dict def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Any: a : List[Any] = self.num_labels a : Any = CTRLForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : Any = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class __UpperCamelCase ( a__ , a__ , a__ , unittest.TestCase ): lowerCamelCase : List[str] =(CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () lowerCamelCase : List[str] =(CTRLLMHeadModel,) if is_torch_available() else () lowerCamelCase : Optional[Any] =( { """feature-extraction""": CTRLModel, """text-classification""": CTRLForSequenceClassification, """text-generation""": CTRLLMHeadModel, """zero-shot""": CTRLForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase : Optional[Any] =True lowerCamelCase : Dict =False lowerCamelCase : Optional[int] =False def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def __a ( self ) -> Union[str, Any]: a : Any = CTRLModelTester(self ) a : Union[str, Any] = ConfigTester(self , config_class=lowerCAmelCase__ , n_embd=37 ) def __a ( self ) -> List[Any]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def __a ( self ) -> List[Any]: a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*lowerCAmelCase__ ) def __a ( self ) -> Any: a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowerCAmelCase__ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> Dict: pass @slow def __a ( self ) -> Dict: for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Tuple = CTRLModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @unittest.skip("The model doesn't support left padding" ) # and it's not used enough to be worth fixing :) def __a ( self ) -> Optional[Any]: pass @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Any: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def __a ( self ) -> Dict: a : Union[str, Any] = CTRLLMHeadModel.from_pretrained("ctrl" ) model.to(lowerCAmelCase__ ) a : str = torch.tensor( [[1_1859, 0, 1611, 8]] , dtype=torch.long , device=lowerCAmelCase__ ) # Legal the president is a : Optional[int] = [ 1_1859, 0, 1611, 8, 5, 150, 2_6449, 2, 19, 348, 469, 3, 2595, 48, 2_0740, 24_6533, 24_6533, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a a : Optional[int] = model.generate(lowerCAmelCase__ , do_sample=lowerCAmelCase__ ) self.assertListEqual(output_ids[0].tolist() , lowerCAmelCase__ )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : Optional[Any] = { '''hustvl/yolos-small''': '''https://huggingface.co/hustvl/yolos-small/resolve/main/config.json''', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class __UpperCamelCase ( a__ ): lowerCamelCase : int ="""yolos""" def __init__( self , lowerCAmelCase__=768 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=3072 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=[512, 864] , lowerCAmelCase__=16 , lowerCAmelCase__=3 , lowerCAmelCase__=True , lowerCAmelCase__=100 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=1 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=0.1 , **lowerCAmelCase__ , ) -> Union[str, Any]: super().__init__(**lowerCAmelCase__ ) a : List[Any] = hidden_size a : Optional[Any] = num_hidden_layers a : List[Any] = num_attention_heads a : int = intermediate_size a : int = hidden_act a : int = hidden_dropout_prob a : Optional[int] = attention_probs_dropout_prob a : Tuple = initializer_range a : Optional[Any] = layer_norm_eps a : Any = image_size a : Optional[int] = patch_size a : str = num_channels a : Dict = qkv_bias a : Union[str, Any] = num_detection_tokens a : Optional[Any] = use_mid_position_embeddings a : List[Any] = auxiliary_loss # Hungarian matcher a : int = class_cost a : str = bbox_cost a : List[Any] = giou_cost # Loss coefficients a : List[str] = bbox_loss_coefficient a : Any = giou_loss_coefficient a : Optional[Any] = eos_coefficient class __UpperCamelCase ( a__ ): lowerCamelCase : Tuple =version.parse("""1.11""" ) @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def __a ( self ) -> float: return 1E-4 @property def __a ( self ) -> int: return 12
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ = "▁" , lowerCAmelCase__ = True , lowerCAmelCase__ = "<unk>" , lowerCAmelCase__ = "</s>" , lowerCAmelCase__ = "<pad>" , ) -> Optional[Any]: a : List[str] = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } a : str = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): a : Dict = token_dict["token"] a : List[str] = Tokenizer(Unigram() ) a : int = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ) , " " ), normalizers.Lowercase(), ] ) a : str = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ ), pre_tokenizers.Digits(individual_digits=lowerCAmelCase__ ), pre_tokenizers.Punctuation(), ] ) a : Union[str, Any] = decoders.Metaspace(replacement=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ ) a : int = TemplateProcessing( single=f"""$A {self.special_tokens['eos']['token']}""" , special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])] , ) a : Any = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = 8000 , lowerCAmelCase__ = True , ) -> Optional[int]: a : str = trainers.UnigramTrainer( vocab_size=lowerCAmelCase__ , special_tokens=self.special_tokens_list , show_progress=lowerCAmelCase__ , ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] self._tokenizer.train(lowerCAmelCase__ , trainer=lowerCAmelCase__ ) self.add_unk_id() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = 8000 , lowerCAmelCase__ = True , ) -> List[str]: a : List[Any] = trainers.UnigramTrainer( vocab_size=lowerCAmelCase__ , special_tokens=self.special_tokens_list , show_progress=lowerCAmelCase__ , ) self._tokenizer.train_from_iterator(lowerCAmelCase__ , trainer=lowerCAmelCase__ ) self.add_unk_id() def __a ( self ) -> Tuple: a : List[Any] = json.loads(self._tokenizer.to_str() ) a : Dict = self.special_tokens["unk"]["id"] a : str = Tokenizer.from_str(json.dumps(lowerCAmelCase__ ) )
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging a : Optional[int] = logging.get_logger(__name__) a : Union[str, Any] = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED a : int = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } a : Optional[Any] = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def _SCREAMING_SNAKE_CASE ( ) ->int: '''simple docstring''' a : Dict = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) a : List[Any] = bs[:] a : Union[str, Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowercase ) cs.append(2**8 + n ) n += 1 a : Union[str, Any] = [chr(_lowercase ) for n in cs] return dict(zip(_lowercase , _lowercase ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple ) ->int: '''simple docstring''' a : List[str] = set() a : Optional[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) a : List[str] = char return pairs class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : Union[str, Any] =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Optional[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : str =["""input_ids""", """attention_mask"""] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="replace" , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: a : Any = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else bos_token a : str = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else eos_token a : List[str] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else sep_token a : List[Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else cls_token a : List[Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else unk_token a : int = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it a : Optional[Any] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ , **lowerCAmelCase__ , ) with open(lowerCAmelCase__ , encoding="utf-8" ) as vocab_handle: a : Optional[int] = json.load(lowerCAmelCase__ ) a : Any = {v: k for k, v in self.encoder.items()} a : Tuple = errors # how to handle errors in decoding a : List[Any] = bytes_to_unicode() a : int = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ , encoding="utf-8" ) as merges_handle: a : Tuple = merges_handle.read().split("\n" )[1:-1] a : str = [tuple(merge.split() ) for merge in bpe_merges] a : Tuple = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) a : int = {} a : int = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions a : Tuple = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def __a ( self ) -> int: return len(self.encoder ) def __a ( self ) -> List[Any]: return dict(self.encoder , **self.added_tokens_encoder ) def __a ( self , lowerCAmelCase__ ) -> List[str]: if token in self.cache: return self.cache[token] a : str = tuple(lowerCAmelCase__ ) a : Dict = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: a : Tuple = min(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break a, a : List[str] = bigram a : List[Any] = [] a : List[Any] = 0 while i < len(lowerCAmelCase__ ): try: a : str = word.index(lowerCAmelCase__ , lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) a : Tuple = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 a : Union[str, Any] = tuple(lowerCAmelCase__ ) a : str = new_word if len(lowerCAmelCase__ ) == 1: break else: a : List[Any] = get_pairs(lowerCAmelCase__ ) a : Any = " ".join(lowerCAmelCase__ ) a : Optional[Any] = word return word def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: a : Union[str, Any] = [] for token in re.findall(self.pat , lowerCAmelCase__ ): a : List[Any] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def __a ( self , lowerCAmelCase__ ) -> Dict: return self.encoder.get(lowerCAmelCase__ , self.encoder.get(self.unk_token ) ) def __a ( self , lowerCAmelCase__ ) -> str: return self.decoder.get(lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ ) -> List[str]: a : Union[str, Any] = "".join(lowerCAmelCase__ ) a : Any = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : str = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) a : Tuple = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ ) + "\n" ) a : List[Any] = 0 with open(lowerCAmelCase__ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) a : Union[str, Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] a : Optional[Any] = [self.cls_token_id] a : Any = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[int] = [self.sep_token_id] a : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=False , **lowerCAmelCase__ ) -> List[str]: a : List[Any] = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): a : Tuple = " " + text return (text, kwargs) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = PaddingStrategy.DO_NOT_PAD , lowerCAmelCase__ = None , lowerCAmelCase__ = None , ) -> dict: a : Dict = super()._pad( encoded_inputs=lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding_strategy=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , ) # Load from model defaults if return_attention_mask is None: a : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: a : List[Any] = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. a : Tuple = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: a : List[str] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` a : Union[str, Any] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": a : str = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __UpperCamelCase ( a__ ): lowerCamelCase : int =["""image_processor""", """tokenizer"""] lowerCamelCase : str ="""BlipImageProcessor""" lowerCamelCase : int ="""AutoTokenizer""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Tuple = False super().__init__(lowerCAmelCase__ , lowerCAmelCase__ ) a : Union[str, Any] = self.image_processor def __call__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 0 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = True , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> BatchEncoding: if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: a : int = self.tokenizer a : List[str] = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) return text_encoding # add pixel_values a : str = self.image_processor(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) if text is not None: a : List[str] = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) else: a : Optional[int] = None if text_encoding is not None: encoding_image_processor.update(lowerCAmelCase__ ) return encoding_image_processor def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: return self.tokenizer.batch_decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[Any]: return self.tokenizer.decode(*lowerCAmelCase__ , **lowerCAmelCase__ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __a ( self ) -> Any: a : List[Any] = self.tokenizer.model_input_names a : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: if dst_width < 0 or dst_height < 0: raise ValueError("Destination width/height should be > 0" ) a : str = img a : Any = img.shape[1] a : Union[str, Any] = img.shape[0] a : str = dst_width a : Optional[int] = dst_height a : Union[str, Any] = self.src_w / self.dst_w a : List[str] = self.src_h / self.dst_h a : Tuple = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 255 ) def __a ( self ) -> Dict: for i in range(self.dst_h ): for j in range(self.dst_w ): a : Any = self.img[self.get_y(lowerCAmelCase__ )][self.get_x(lowerCAmelCase__ )] def __a ( self , lowerCAmelCase__ ) -> int: return int(self.ratio_x * x ) def __a ( self , lowerCAmelCase__ ) -> int: return int(self.ratio_y * y ) if __name__ == "__main__": a , a : Any = 800, 600 a : Union[str, Any] = imread('''image_data/lena.jpg''', 1) a : List[Any] = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F'''Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}''', n.output ) waitKey(0) destroyAllWindows()
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate a : Dict = trt.Logger(trt.Logger.WARNING) a : Optional[Any] = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) a : str = logging.getLogger(__name__) a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--onnx_model_path''', default=None, type=str, required=True, help='''Path to ONNX model: ''', ) parser.add_argument( '''--output_dir''', default=None, type=str, required=True, help='''The output directory where the model checkpoints and predictions will be written.''', ) # Other parameters parser.add_argument( '''--tokenizer_name''', default='''''', type=str, required=True, help='''Pretrained tokenizer name or path if not the same as model_name''', ) parser.add_argument( '''--version_2_with_negative''', action='''store_true''', help='''If true, the SQuAD examples contain some that do not have an answer.''', ) parser.add_argument( '''--null_score_diff_threshold''', type=float, default=0.0, help='''If null_score - best_non_null is greater than the threshold predict null.''', ) parser.add_argument( '''--max_seq_length''', default=384, type=int, help=( '''The maximum total input sequence length after WordPiece tokenization. Sequences ''' '''longer than this will be truncated, and sequences shorter than this will be padded.''' ), ) parser.add_argument( '''--doc_stride''', default=128, type=int, help='''When splitting up a long document into chunks, how much stride to take between chunks.''', ) parser.add_argument('''--per_device_eval_batch_size''', default=8, type=int, help='''Batch size per GPU/CPU for evaluation.''') parser.add_argument( '''--n_best_size''', default=20, type=int, help='''The total number of n-best predictions to generate in the nbest_predictions.json output file.''', ) parser.add_argument( '''--max_answer_length''', default=30, type=int, help=( '''The maximum length of an answer that can be generated. This is needed because the start ''' '''and end predictions are not conditioned on one another.''' ), ) parser.add_argument('''--seed''', type=int, default=42, help='''random seed for initialization''') parser.add_argument( '''--dataset_name''', type=str, default=None, required=True, help='''The name of the dataset to use (via the datasets library).''', ) parser.add_argument( '''--dataset_config_name''', type=str, default=None, help='''The configuration name of the dataset to use (via the datasets library).''', ) parser.add_argument( '''--preprocessing_num_workers''', type=int, default=4, help='''A csv or a json file containing the training data.''' ) parser.add_argument('''--overwrite_cache''', action='''store_true''', help='''Overwrite the cached training and evaluation sets''') parser.add_argument( '''--fp16''', action='''store_true''', help='''Whether to use 16-bit (mixed) precision instead of 32-bit''', ) parser.add_argument( '''--int8''', action='''store_true''', help='''Whether to use INT8''', ) a : Optional[int] = parser.parse_args() if args.tokenizer_name: a : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported by this script.''' '''You can do it from another script, save it, and load it from here, using --tokenizer_name.''' ) logger.info('''Training/evaluation parameters %s''', args) a : Tuple = args.per_device_eval_batch_size a : Dict = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties a : Dict = True a : List[Any] = '''temp_engine/bert-fp32.engine''' if args.fpaa: a : Optional[int] = '''temp_engine/bert-fp16.engine''' if args.inta: a : List[str] = '''temp_engine/bert-int8.engine''' # import ONNX file if not os.path.exists('''temp_engine'''): os.makedirs('''temp_engine''') a : Optional[Any] = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, '''rb''') as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network a : List[str] = [network.get_input(i) for i in range(network.num_inputs)] a : str = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: a : str = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) a : Optional[int] = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) a : int = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, '''wb''') as f: f.write(engine.serialize()) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : Optional[int] , _lowercase : List[Any] , _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] , _lowercase : str , _lowercase : Tuple ) ->int: '''simple docstring''' a : List[Any] = np.asarray(inputs["input_ids"] , dtype=np.intaa ) a : Any = np.asarray(inputs["attention_mask"] , dtype=np.intaa ) a : Optional[Any] = np.asarray(inputs["token_type_ids"] , dtype=np.intaa ) # Copy inputs cuda.memcpy_htod_async(d_inputs[0] , input_ids.ravel() , _lowercase ) cuda.memcpy_htod_async(d_inputs[1] , attention_mask.ravel() , _lowercase ) cuda.memcpy_htod_async(d_inputs[2] , token_type_ids.ravel() , _lowercase ) # start time a : Optional[int] = time.time() # Run inference context.execute_async( bindings=[int(_lowercase ) for d_inp in d_inputs] + [int(_lowercase ), int(_lowercase )] , stream_handle=stream.handle ) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(_lowercase , _lowercase , _lowercase ) cuda.memcpy_dtoh_async(_lowercase , _lowercase , _lowercase ) # Synchronize the stream and take time stream.synchronize() # end time a : str = time.time() a : Any = end_time - start_time a : List[Any] = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. a : int = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. a : int = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError('''Evaluation requires a dataset name''') # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. a : Dict = raw_datasets['''validation'''].column_names a : Tuple = '''question''' if '''question''' in column_names else column_names[0] a : Dict = '''context''' if '''context''' in column_names else column_names[1] a : Any = '''answers''' if '''answers''' in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). a : Any = tokenizer.padding_side == '''right''' if args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) a : int = min(args.max_seq_length, tokenizer.model_max_length) def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Union[str, Any]: '''simple docstring''' a : List[Any] = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. a : Optional[int] = tokenizer( examples[question_column_name if pad_on_right else context_column_name] , examples[context_column_name if pad_on_right else question_column_name] , truncation="only_second" if pad_on_right else "only_first" , max_length=_lowercase , stride=args.doc_stride , return_overflowing_tokens=_lowercase , return_offsets_mapping=_lowercase , padding="max_length" , ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. a : List[str] = tokenized_examples.pop("overflow_to_sample_mapping" ) # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. a : str = [] for i in range(len(tokenized_examples["input_ids"] ) ): # Grab the sequence corresponding to that example (to know what is the context and what is the question). a : Tuple = tokenized_examples.sequence_ids(_lowercase ) a : Optional[Any] = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. a : Dict = sample_mapping[i] tokenized_examples["example_id"].append(examples["id"][sample_index] ) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. a : Any = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples["offset_mapping"][i] ) ] return tokenized_examples a : Union[str, Any] = raw_datasets['''validation'''] # Validation Feature Creation a : Any = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc='''Running tokenizer on validation dataset''', ) a : Optional[int] = default_data_collator a : Tuple = eval_dataset.remove_columns(['''example_id''', '''offset_mapping''']) a : Optional[Any] = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : str , _lowercase : int , _lowercase : int="eval" ) ->List[str]: '''simple docstring''' a : Union[str, Any] = postprocess_qa_predictions( examples=_lowercase , features=_lowercase , predictions=_lowercase , version_2_with_negative=args.version_2_with_negative , n_best_size=args.n_best_size , max_answer_length=args.max_answer_length , null_score_diff_threshold=args.null_score_diff_threshold , output_dir=args.output_dir , prefix=_lowercase , ) # Format the result to the format the metric expects. if args.version_2_with_negative: a : Optional[Any] = [ {"id": k, "prediction_text": v, "no_answer_probability": 0.0} for k, v in predictions.items() ] else: a : List[str] = [{"id": k, "prediction_text": v} for k, v in predictions.items()] a : Optional[Any] = [{"id": ex["id"], "answers": ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=_lowercase , label_ids=_lowercase ) a : Union[str, Any] = load_metric('''squad_v2''' if args.version_2_with_negative else '''squad''') # Evaluation! logger.info('''Loading ONNX model %s for evaluation''', args.onnx_model_path) with open(engine_name, '''rb''') as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->List[Any]: '''simple docstring''' return trt.volume(engine.get_binding_shape(_lowercase ) ) * engine.get_binding_dtype(_lowercase ).itemsize # Allocate device memory for inputs and outputs. a : Dict = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer a : str = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) a : Union[str, Any] = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) a : List[Any] = cuda.mem_alloc(h_outputa.nbytes) a : List[str] = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. a : int = cuda.Stream() # Evaluation logger.info('''***** Running Evaluation *****''') logger.info(F''' Num examples = {len(eval_dataset)}''') logger.info(F''' Batch size = {args.per_device_eval_batch_size}''') a : str = 0.0 a : List[Any] = 0 a : Any = timeit.default_timer() a : Optional[int] = None for step, batch in enumerate(eval_dataloader): a , a : Optional[int] = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 a , a : Any = outputs a : Tuple = torch.tensor(start_logits) a : Union[str, Any] = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered a : int = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-100) a : List[str] = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-100) a : Tuple = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) a : str = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-100) if all_preds is not None: a : Any = nested_truncate(all_preds, len(eval_dataset)) a : Union[str, Any] = timeit.default_timer() - start_time logger.info(''' Evaluation done in total %f secs (%f sec per example)''', evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info('''Average Inference Time = {:.3f} ms'''.format(total_time * 1000 / niter)) logger.info('''Total Inference Time = {:.3f} ms'''.format(total_time * 1000)) logger.info('''Total Number of Inference = %d''', niter) a : Union[str, Any] = post_processing_function(eval_examples, eval_dataset, all_preds) a : int = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(F'''Evaluation metrics: {eval_metric}''')
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : List[Any] ) ->Optional[Any]: '''simple docstring''' assert isinstance(_lowercase , _lowercase ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : int , _lowercase : Optional[Any] ) ->List[Any]: '''simple docstring''' a : Optional[int] = tmp_path / "cache" a : List[Any] = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a : Dict = TextDatasetReader(_lowercase , cache_dir=_lowercase , keep_in_memory=_lowercase ).read() _check_text_dataset(_lowercase , _lowercase ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict , _lowercase : str , _lowercase : Dict ) ->Union[str, Any]: '''simple docstring''' a : Tuple = tmp_path / "cache" a : List[Any] = {"text": "string"} a : Tuple = features.copy() if features else default_expected_features a : List[str] = ( Features({feature: Value(_lowercase ) for feature, dtype in features.items()} ) if features is not None else None ) a : Any = TextDatasetReader(_lowercase , features=_lowercase , cache_dir=_lowercase ).read() _check_text_dataset(_lowercase , _lowercase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : str , _lowercase : List[str] ) ->str: '''simple docstring''' a : List[Any] = tmp_path / "cache" a : Optional[int] = {"text": "string"} a : Tuple = TextDatasetReader(_lowercase , cache_dir=_lowercase , split=_lowercase ).read() _check_text_dataset(_lowercase , _lowercase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Tuple , _lowercase : List[Any] ) ->List[Any]: '''simple docstring''' if issubclass(_lowercase , _lowercase ): a : Any = text_path elif issubclass(_lowercase , _lowercase ): a : Optional[Any] = [text_path] a : int = tmp_path / "cache" a : int = {"text": "string"} a : Dict = TextDatasetReader(_lowercase , cache_dir=_lowercase ).read() _check_text_dataset(_lowercase , _lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Dict , _lowercase : Any=("train",) ) ->List[str]: '''simple docstring''' assert isinstance(_lowercase , _lowercase ) for split in splits: a : int = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : List[str] , _lowercase : Union[str, Any] ) ->str: '''simple docstring''' a : Dict = tmp_path / "cache" a : List[str] = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a : Dict = TextDatasetReader({"train": text_path} , cache_dir=_lowercase , keep_in_memory=_lowercase ).read() _check_text_datasetdict(_lowercase , _lowercase ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : Dict , _lowercase : str ) ->Dict: '''simple docstring''' a : List[str] = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" a : Any = {"text": "string"} a : Union[str, Any] = features.copy() if features else default_expected_features a : List[Any] = ( Features({feature: Value(_lowercase ) for feature, dtype in features.items()} ) if features is not None else None ) a : List[str] = TextDatasetReader({"train": text_path} , features=_lowercase , cache_dir=_lowercase ).read() _check_text_datasetdict(_lowercase , _lowercase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Any , _lowercase : str ) ->Union[str, Any]: '''simple docstring''' if split: a : Optional[Any] = {split: text_path} else: a : Any = "train" a : int = {"train": text_path, "test": text_path} a : Optional[int] = tmp_path / "cache" a : Any = {"text": "string"} a : Optional[Any] = TextDatasetReader(_lowercase , cache_dir=_lowercase ).read() _check_text_datasetdict(_lowercase , _lowercase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : Tuple = {'''configuration_mbart''': ['''MBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MBartConfig''', '''MBartOnnxConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : int = ['''MBartTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[int] = ['''MBartTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = [ '''MBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MBartForCausalLM''', '''MBartForConditionalGeneration''', '''MBartForQuestionAnswering''', '''MBartForSequenceClassification''', '''MBartModel''', '''MBartPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = [ '''TFMBartForConditionalGeneration''', '''TFMBartModel''', '''TFMBartPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''FlaxMBartForConditionalGeneration''', '''FlaxMBartForQuestionAnswering''', '''FlaxMBartForSequenceClassification''', '''FlaxMBartModel''', '''FlaxMBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys a : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer a : str = logging.get_logger(__name__) # pylint: disable=invalid-name a : Any = ''' Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu") >>> repo = "openai/shap-e-img2img" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png" >>> image = load_image(image_url).convert("RGB") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], "corgi_3d.gif") ``` ''' @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : Union[PIL.Image.Image, np.ndarray] class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> Optional[int]: super().__init__() self.register_modules( prior=lowerCAmelCase__ , image_encoder=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , renderer=lowerCAmelCase__ , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: if latents is None: a : List[Any] = randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) a : Optional[int] = latents.to(lowerCAmelCase__ ) a : Dict = latents * scheduler.init_noise_sigma return latents def __a ( self , lowerCAmelCase__=0 ) -> List[str]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) a : Tuple = torch.device(f"""cuda:{gpu_id}""" ) a : Union[str, Any] = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCAmelCase__ , lowerCAmelCase__ ) @property def __a ( self ) -> int: if self.device != torch.device("meta" ) or not hasattr(self.image_encoder , "_hf_hook" ): return self.device for module in self.image_encoder.modules(): if ( hasattr(lowerCAmelCase__ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> str: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and isinstance(image[0] , torch.Tensor ): a : Tuple = torch.cat(lowerCAmelCase__ , axis=0 ) if image[0].ndim == 4 else torch.stack(lowerCAmelCase__ , axis=0 ) if not isinstance(lowerCAmelCase__ , torch.Tensor ): a : Union[str, Any] = self.image_processor(lowerCAmelCase__ , return_tensors="pt" ).pixel_values[0].unsqueeze(0 ) a : List[Any] = image.to(dtype=self.image_encoder.dtype , device=lowerCAmelCase__ ) a : Dict = self.image_encoder(lowerCAmelCase__ )["last_hidden_state"] a : Union[str, Any] = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 a : Any = image_embeds.repeat_interleave(lowerCAmelCase__ , dim=0 ) if do_classifier_free_guidance: a : str = torch.zeros_like(lowerCAmelCase__ ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes a : Dict = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(lowerCAmelCase__ ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 25 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 4.0 , lowerCAmelCase__ = 64 , lowerCAmelCase__ = "pil" , lowerCAmelCase__ = True , ) -> List[Any]: if isinstance(lowerCAmelCase__ , PIL.Image.Image ): a : str = 1 elif isinstance(lowerCAmelCase__ , torch.Tensor ): a : str = image.shape[0] elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): a : Union[str, Any] = len(lowerCAmelCase__ ) else: raise ValueError( f"""`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(lowerCAmelCase__ )}""" ) a : Tuple = self._execution_device a : Optional[Any] = batch_size * num_images_per_prompt a : int = guidance_scale > 1.0 a : int = self._encode_image(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # prior self.scheduler.set_timesteps(lowerCAmelCase__ , device=lowerCAmelCase__ ) a : Any = self.scheduler.timesteps a : str = self.prior.config.num_embeddings a : Tuple = self.prior.config.embedding_dim a : str = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim a : Tuple = latents.reshape(latents.shape[0] , lowerCAmelCase__ , lowerCAmelCase__ ) for i, t in enumerate(self.progress_bar(lowerCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance a : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a : int = self.scheduler.scale_model_input(lowerCAmelCase__ , lowerCAmelCase__ ) a : str = self.prior( lowerCAmelCase__ , timestep=lowerCAmelCase__ , proj_embedding=lowerCAmelCase__ , ).predicted_image_embedding # remove the variance a, a : List[Any] = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: a, a : Optional[Any] = noise_pred.chunk(2 ) a : Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) a : List[str] = self.scheduler.step( lowerCAmelCase__ , timestep=lowerCAmelCase__ , sample=lowerCAmelCase__ , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=lowerCAmelCase__ ) a : Any = [] for i, latent in enumerate(lowerCAmelCase__ ): print() a : Dict = self.renderer.decode( latent[None, :] , lowerCAmelCase__ , size=lowerCAmelCase__ , ray_batch_size=4096 , n_coarse_samples=64 , n_fine_samples=128 , ) images.append(lowerCAmelCase__ ) a : int = torch.stack(lowerCAmelCase__ ) if output_type not in ["np", "pil"]: raise ValueError(f"""Only the output types `pil` and `np` are supported not output_type={output_type}""" ) a : Any = images.cpu().numpy() if output_type == "pil": a : Any = [self.numpy_to_pil(lowerCAmelCase__ ) for image in images] # Offload last model to CPU if hasattr(self , "final_offload_hook" ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=lowerCAmelCase__ )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 400_0000 ) ->int: '''simple docstring''' a : Optional[Any] = [] a, a : int = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(_lowercase ) a, a : Any = b, a + b return sum(_lowercase ) if __name__ == "__main__": print(F'''{solution() = }''')
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __UpperCamelCase ( a__ ): def __a ( self ) -> List[str]: a : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCAmelCase__ , "num_attention_heads" ) ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=32 , lowerCAmelCase__=2 , lowerCAmelCase__=3 , lowerCAmelCase__=640 , lowerCAmelCase__=4 , lowerCAmelCase__="silu" , lowerCAmelCase__=3 , lowerCAmelCase__=32 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.02 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=10 , lowerCAmelCase__=None , ) -> Tuple: a : List[Any] = parent a : Tuple = batch_size a : int = image_size a : Tuple = patch_size a : Tuple = num_channels a : str = last_hidden_size a : List[str] = num_attention_heads a : Optional[int] = hidden_act a : Dict = conv_kernel_size a : Any = output_stride a : Tuple = hidden_dropout_prob a : Union[str, Any] = attention_probs_dropout_prob a : str = classifier_dropout_prob a : List[Any] = use_labels a : Dict = is_training a : int = num_labels a : Tuple = initializer_range a : Optional[int] = scope def __a ( self ) -> Optional[Any]: a : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a : int = None a : List[str] = None if self.use_labels: a : int = ids_tensor([self.batch_size] , self.num_labels ) a : Tuple = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) a : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def __a ( self ) -> List[Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: a : Tuple = MobileViTModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Any = model(lowerCAmelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: a : Optional[int] = self.num_labels a : Optional[Any] = MobileViTForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : int = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: a : List[str] = self.num_labels a : List[Any] = MobileViTForSemanticSegmentation(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Optional[Any] = model(lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) a : List[str] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __a ( self ) -> str: a : Tuple = self.prepare_config_and_inputs() a, a, a, a : Any = config_and_inputs a : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : str =( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) lowerCamelCase : List[Any] =( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) lowerCamelCase : Optional[int] =False lowerCamelCase : List[str] =False lowerCamelCase : List[str] =False lowerCamelCase : Any =False def __a ( self ) -> Any: a : Optional[Any] = MobileViTModelTester(self ) a : Union[str, Any] = MobileViTConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __a ( self ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __a ( self ) -> int: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __a ( self ) -> List[str]: pass def __a ( self ) -> Any: a, a : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Union[str, Any] = model_class(lowerCAmelCase__ ) a : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : int = [*signature.parameters.keys()] a : Union[str, Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> Union[str, Any]: pass def __a ( self ) -> List[Any]: a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def __a ( self ) -> str: def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): a : Optional[int] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): a : int = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Any = outputs.hidden_states a : int = 5 self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. a : int = 2 for i in range(len(lowerCAmelCase__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) a, a : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Dict = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a : Optional[Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase__ ) @slow def __a ( self ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : str = MobileViTModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->Optional[Any]: '''simple docstring''' a : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def __a ( self ) -> str: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __a ( self ) -> int: a : List[str] = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCAmelCase__ ) a : Any = self.default_image_processor a : Dict = prepare_img() a : List[str] = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : List[str] = model(**lowerCAmelCase__ ) # verify the logits a : Any = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) a : Optional[Any] = torch.tensor([-1.9_364, -1.2_327, -0.4_653] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def __a ( self ) -> str: a : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) a : int = model.to(lowerCAmelCase__ ) a : Union[str, Any] = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) a : Optional[int] = prepare_img() a : Any = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : Tuple = model(**lowerCAmelCase__ ) a : Union[str, Any] = outputs.logits # verify the logits a : Dict = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCAmelCase__ ) a : List[str] = torch.tensor( [ [[6.9_713, 6.9_786, 7.2_422], [7.2_893, 7.2_825, 7.4_446], [7.6_580, 7.8_797, 7.9_420]], [[-10.6_869, -10.3_250, -10.3_471], [-10.4_228, -9.9_868, -9.7_132], [-11.0_405, -11.0_221, -10.7_318]], [[-3.3_089, -2.8_539, -2.6_740], [-3.2_706, -2.5_621, -2.5_108], [-3.2_534, -2.6_615, -2.6_651]], ] , device=lowerCAmelCase__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def __a ( self ) -> str: a : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) a : Dict = model.to(lowerCAmelCase__ ) a : Optional[Any] = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) a : Union[str, Any] = prepare_img() a : Tuple = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): a : Tuple = model(**lowerCAmelCase__ ) a : Union[str, Any] = outputs.logits.detach().cpu() a : int = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__ , target_sizes=[(50, 60)] ) a : Tuple = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCAmelCase__ ) a : Tuple = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__ ) a : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCAmelCase__ )
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ) -> List[Any]: a : Dict = parent a : List[Any] = batch_size a : int = seq_length a : Optional[Any] = is_training a : List[str] = use_token_type_ids a : int = use_labels a : Dict = vocab_size a : Any = hidden_size a : List[Any] = num_hidden_layers a : List[Any] = num_attention_heads a : Optional[int] = intermediate_size a : List[Any] = hidden_act a : List[Any] = hidden_dropout_prob a : Dict = attention_probs_dropout_prob a : Tuple = max_position_embeddings a : Tuple = type_vocab_size a : int = type_sequence_label_size a : str = initializer_range a : str = num_labels a : int = num_choices a : str = scope a : Union[str, Any] = self.vocab_size - 1 def __a ( self ) -> int: a : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : str = None if self.use_token_type_ids: a : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a : List[Any] = None a : int = None a : Optional[int] = None if self.use_labels: a : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a : Any = ids_tensor([self.batch_size] , self.num_choices ) a : Any = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) a : Optional[Any] = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Union[str, Any]: a : Tuple = OpenAIGPTModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : List[str] = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , head_mask=lowerCAmelCase__ ) a : Optional[int] = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ ) a : Optional[int] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Tuple: a : Any = OpenAIGPTLMHeadModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : List[str] = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Tuple: a : Tuple = OpenAIGPTDoubleHeadsModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : Dict = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ ) -> Dict: a : Optional[Any] = self.num_labels a : Dict = OpenAIGPTForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() a : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : Tuple = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self ) -> Optional[int]: a : Optional[int] = self.prepare_config_and_inputs() ( ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ( a ), ) : Tuple = config_and_inputs a : List[str] = { "input_ids": input_ids, "token_type_ids": token_type_ids, "head_mask": head_mask, } return config, inputs_dict @require_torch class __UpperCamelCase ( a__ , a__ , a__ , unittest.TestCase ): lowerCamelCase : List[Any] =( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) lowerCamelCase : Dict =( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly lowerCamelCase : Any =( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ) -> Tuple: a : Dict = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": a : Tuple = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=lowerCAmelCase__ , ) a : Optional[Any] = inputs_dict["labels"] a : Union[str, Any] = inputs_dict["labels"] a : int = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=lowerCAmelCase__ , ) a : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase__ ) return inputs_dict def __a ( self ) -> Dict: a : List[str] = OpenAIGPTModelTester(self ) a : Any = ConfigTester(self , config_class=lowerCAmelCase__ , n_embd=37 ) def __a ( self ) -> Dict: self.config_tester.run_common_tests() def __a ( self ) -> str: a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowerCAmelCase__ ) def __a ( self ) -> List[str]: a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowerCAmelCase__ ) def __a ( self ) -> Any: a : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowerCAmelCase__ ) @slow def __a ( self ) -> Dict: for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : List[Any] = OpenAIGPTModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Any: a : str = OpenAIGPTLMHeadModel.from_pretrained("openai-gpt" ) model.to(lowerCAmelCase__ ) a : Union[str, Any] = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=lowerCAmelCase__ ) # the president is a : str = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 4_0477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the a : Optional[Any] = model.generate(lowerCAmelCase__ , do_sample=lowerCAmelCase__ ) self.assertListEqual(output_ids[0].tolist() , lowerCAmelCase__ )
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" import argparse from collections import defaultdict def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : List[Any] , _lowercase : str , _lowercase : Optional[Any] , _lowercase : List[Any] ) ->int: '''simple docstring''' a : List[str] = F"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_lowercase , "r" ) as f: a : str = f.readlines() a : Any = F"""class {class_name}(""" a : Any = F"""{4 * ' '}def {test_name}(""" a : Optional[Any] = F"""{8 * ' '}{correct_line.split()[0]}""" a : Any = F"""{16 * ' '}{correct_line.split()[0]}""" a : Any = False a : str = False a : Union[str, Any] = False a : Optional[int] = False a : List[str] = 0 a : Optional[int] = 0 a : List[Any] = [] for line in lines: if line.startswith(_lowercase ): a : Tuple = True elif in_class and line.startswith(_lowercase ): a : Optional[Any] = True elif in_class and in_func and (line.startswith(_lowercase ) or line.startswith(_lowercase )): a : List[Any] = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: a : Optional[Any] = True if in_class and in_func and in_line: if ")" not in line: continue else: a : str = True if in_class and in_func and in_line and insert_line: new_lines.append(F"""{spaces * ' '}{correct_line}""" ) a : Optional[Any] = False else: new_lines.append(_lowercase ) with open(_lowercase , "w" ) as f: for line in new_lines: f.write(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] , _lowercase : List[str]=None ) ->Dict: '''simple docstring''' if fail is not None: with open(_lowercase , "r" ) as f: a : Optional[int] = {l.strip() for l in f.readlines()} else: a : Optional[Any] = None with open(_lowercase , "r" ) as f: a : Any = f.readlines() a : List[Any] = defaultdict(_lowercase ) for line in correct_lines: a, a, a, a : Any = line.split(";" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) if __name__ == "__main__": a : str = argparse.ArgumentParser() parser.add_argument('''--correct_filename''', help='''filename of tests with expected result''') parser.add_argument('''--fail_filename''', help='''filename of test failures''', type=str, default=None) a : Tuple = parser.parse_args() main(args.correct_filename, args.fail_filename)
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : Optional[int] = logging.get_logger(__name__) a : Optional[int] = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __UpperCamelCase ( a__ ): lowerCamelCase : str ="""swinv2""" lowerCamelCase : Any ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , lowerCAmelCase__=224 , lowerCAmelCase__=4 , lowerCAmelCase__=3 , lowerCAmelCase__=96 , lowerCAmelCase__=[2, 2, 6, 2] , lowerCAmelCase__=[3, 6, 12, 24] , lowerCAmelCase__=7 , lowerCAmelCase__=4.0 , lowerCAmelCase__=True , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__="gelu" , lowerCAmelCase__=False , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=32 , **lowerCAmelCase__ , ) -> Optional[int]: super().__init__(**lowerCAmelCase__ ) a : Optional[Any] = image_size a : Any = patch_size a : int = num_channels a : Any = embed_dim a : Union[str, Any] = depths a : Dict = len(lowerCAmelCase__ ) a : int = num_heads a : int = window_size a : List[str] = mlp_ratio a : Dict = qkv_bias a : Tuple = hidden_dropout_prob a : str = attention_probs_dropout_prob a : Any = drop_path_rate a : int = hidden_act a : List[str] = use_absolute_embeddings a : str = layer_norm_eps a : int = initializer_range a : Tuple = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model a : Any = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) a : Dict = (0, 0, 0, 0)
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class __UpperCamelCase : lowerCamelCase : Tuple =None def __a ( self ) -> Dict: a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Dict = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: a : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: a : Tuple = os.path.join(lowerCAmelCase__ , "feat_extract.json" ) feat_extract_first.to_json_file(lowerCAmelCase__ ) a : Dict = self.feature_extraction_class.from_json_file(lowerCAmelCase__ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def __a ( self ) -> int: a : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: a : int = feat_extract_first.save_pretrained(lowerCAmelCase__ )[0] check_json_file_has_correct_format(lowerCAmelCase__ ) a : str = self.feature_extraction_class.from_pretrained(lowerCAmelCase__ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def __a ( self ) -> Tuple: a : int = self.feature_extraction_class() self.assertIsNotNone(lowerCAmelCase__ )
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html a : List[str] = '''platform''' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : List[str] , _lowercase : Optional[Any]=None , _lowercase : List[Any]=None , _lowercase : List[str]=None , _lowercase : Optional[int]=None , _lowercase : Tuple=None , _lowercase : str=None , ) ->int: '''simple docstring''' if attention_mask is None: a : str = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: a : Dict = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: a : Union[str, Any] = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: a : Tuple = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: a : List[Any] = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=99 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=32 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=0.02 , ) -> List[Any]: a : Any = parent a : Optional[Any] = batch_size a : Any = seq_length a : Tuple = is_training a : Optional[Any] = use_labels a : int = vocab_size a : List[Any] = hidden_size a : int = num_hidden_layers a : Optional[Any] = num_attention_heads a : List[str] = intermediate_size a : Optional[int] = hidden_act a : Optional[int] = hidden_dropout_prob a : Tuple = attention_probs_dropout_prob a : Dict = max_position_embeddings a : Optional[Any] = eos_token_id a : Tuple = pad_token_id a : Dict = bos_token_id a : Union[str, Any] = initializer_range def __a ( self ) -> Optional[Any]: a : Dict = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) a : Optional[Any] = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) a : Union[str, Any] = shift_tokens_right(lowerCAmelCase__ , 1 , 2 ) a : List[str] = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCAmelCase__ , ) a : Optional[int] = prepare_blenderbot_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def __a ( self ) -> Optional[int]: a, a : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : Dict = 20 a : Dict = model_class_name(lowerCAmelCase__ ) a : Optional[Any] = model.encode(inputs_dict["input_ids"] ) a, a : Optional[int] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) a : List[Any] = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) a : Tuple = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) a : List[Any] = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) a : Tuple = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) a : Dict = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase__ , ) a : List[Any] = model.decode(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: a : List[Any] = 20 a : Optional[Any] = model_class_name(lowerCAmelCase__ ) a : Optional[int] = model.encode(inputs_dict["input_ids"] ) a, a : Union[str, Any] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) a : List[str] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) a : Union[str, Any] = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) a : int = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) a : List[Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) a : List[Any] = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) a : Union[str, Any] = model.decode(lowerCAmelCase__ , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ ) a : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) @require_flax class __UpperCamelCase ( unittest.TestCase ): lowerCamelCase : int =99 def __a ( self ) -> Union[str, Any]: a : int = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) a : List[Any] = input_ids.shape[0] a : Any = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def __a ( self ) -> Any: a, a, a : str = self._get_config_and_data() a : str = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase__ ) a : List[str] = lm_model(input_ids=lowerCAmelCase__ ) a : Optional[Any] = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCAmelCase__ ) def __a ( self ) -> str: a : Dict = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) a : List[Any] = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase__ ) a : List[Any] = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) a : str = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) a : List[Any] = lm_model(input_ids=lowerCAmelCase__ , decoder_input_ids=lowerCAmelCase__ ) a : List[Any] = (*summary.shape, config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCAmelCase__ ) def __a ( self ) -> List[Any]: a : Optional[int] = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) a : Optional[Any] = shift_tokens_right(lowerCAmelCase__ , 1 , 2 ) a : Dict = np.equal(lowerCAmelCase__ , 1 ).astype(np.floataa ).sum() a : Optional[Any] = np.equal(lowerCAmelCase__ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCAmelCase__ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class __UpperCamelCase ( a__ , unittest.TestCase , a__ ): lowerCamelCase : Optional[int] =True lowerCamelCase : Dict =( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCamelCase : Dict =(FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def __a ( self ) -> Dict: a : Union[str, Any] = FlaxBlenderbotModelTester(self ) def __a ( self ) -> Optional[Any]: a, a : Tuple = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> List[str]: a, a : Dict = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> str: a, a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): a : Any = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) a : int = model_class(lowerCAmelCase__ ) @jax.jit def encode_jitted(lowerCAmelCase__ , lowerCAmelCase__=None , **lowerCAmelCase__ ): return model.encode(input_ids=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) with self.subTest("JIT Enabled" ): a : Union[str, Any] = encode_jitted(**lowerCAmelCase__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): a : Optional[Any] = encode_jitted(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) ) for jitted_output, output in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def __a ( self ) -> str: a, a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): a : Optional[int] = model_class(lowerCAmelCase__ ) a : List[str] = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) a : Dict = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): return model.decode( decoder_input_ids=lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , encoder_outputs=lowerCAmelCase__ , ) with self.subTest("JIT Enabled" ): a : List[str] = decode_jitted(**lowerCAmelCase__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): a : Union[str, Any] = decode_jitted(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) ) for jitted_output, output in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def __a ( self ) -> Tuple: for model_class_name in self.all_model_classes: a : Optional[int] = model_class_name.from_pretrained("facebook/blenderbot-400M-distill" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids a : Union[str, Any] = np.ones((1, 1) ) * model.config.eos_token_id a : str = model(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @unittest.skipUnless(jax_device != "cpu" , "3B test too slow on CPU." ) @slow def __a ( self ) -> List[Any]: a : Optional[Any] = {"num_beams": 1, "early_stopping": True, "min_length": 15, "max_length": 25} a : Tuple = {"skip_special_tokens": True, "clean_up_tokenization_spaces": True} a : Optional[int] = FlaxBlenderbotForConditionalGeneration.from_pretrained("facebook/blenderbot-3B" , from_pt=lowerCAmelCase__ ) a : Tuple = BlenderbotTokenizer.from_pretrained("facebook/blenderbot-3B" ) a : Union[str, Any] = ["Sam"] a : Union[str, Any] = tokenizer(lowerCAmelCase__ , return_tensors="jax" ) a : int = model.generate(**lowerCAmelCase__ , **lowerCAmelCase__ ) a : str = "Sam is a great name. It means \"sun\" in Gaelic." a : Optional[int] = tokenizer.batch_decode(lowerCAmelCase__ , **lowerCAmelCase__ ) assert generated_txt[0].strip() == tgt_text
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __UpperCamelCase ( a__ ): lowerCamelCase : Tuple =(KDPMaDiscreteScheduler,) lowerCamelCase : Any =10 def __a ( self , **lowerCAmelCase__ ) -> Optional[Any]: a : Optional[Any] = { "num_train_timesteps": 1100, "beta_start": 0.0_001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowerCAmelCase__ ) return config def __a ( self ) -> Optional[int]: for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__ ) def __a ( self ) -> Optional[Any]: for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=lowerCAmelCase__ , beta_end=lowerCAmelCase__ ) def __a ( self ) -> Dict: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCAmelCase__ ) def __a ( self ) -> List[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__ ) def __a ( self ) -> Dict: a : Optional[Any] = self.scheduler_classes[0] a : Dict = self.get_scheduler_config(prediction_type="v_prediction" ) a : Dict = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) a : List[str] = self.dummy_model() a : Dict = self.dummy_sample_deter * scheduler.init_noise_sigma a : Dict = sample.to(lowerCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): a : List[str] = scheduler.scale_model_input(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = model(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = scheduler.step(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : Optional[Any] = output.prev_sample a : Tuple = torch.sum(torch.abs(lowerCAmelCase__ ) ) a : List[Any] = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1_112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693_428_650_170_972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_002 ) < 1E-3 def __a ( self ) -> Tuple: if torch_device == "mps": return a : str = self.scheduler_classes[0] a : str = self.get_scheduler_config() a : Dict = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) a : Dict = self.dummy_model() a : List[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma a : Tuple = sample.to(lowerCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): a : List[str] = scheduler.scale_model_input(lowerCAmelCase__ , lowerCAmelCase__ ) a : Union[str, Any] = model(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = scheduler.step(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : int = output.prev_sample a : List[str] = torch.sum(torch.abs(lowerCAmelCase__ ) ) a : List[str] = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 def __a ( self ) -> Dict: if torch_device == "mps": return a : Any = self.scheduler_classes[0] a : Tuple = self.get_scheduler_config() a : List[Any] = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps , device=lowerCAmelCase__ ) a : Optional[Any] = self.dummy_model() a : int = self.dummy_sample_deter.to(lowerCAmelCase__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: a : str = scheduler.scale_model_input(lowerCAmelCase__ , lowerCAmelCase__ ) a : Optional[int] = model(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = scheduler.step(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = output.prev_sample a : Dict = torch.sum(torch.abs(lowerCAmelCase__ ) ) a : List[str] = torch.mean(torch.abs(lowerCAmelCase__ ) ) if str(lowerCAmelCase__ ).startswith("cpu" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class __UpperCamelCase ( a__ ): @slow @require_torch def __a ( self ) -> Union[str, Any]: a : Tuple = EncoderDecoderModel.from_encoder_decoder_pretrained("prajjwal1/bert-tiny" , "prajjwal1/bert-tiny" ) a : List[Any] = BertTokenizer.from_pretrained("bert-base-uncased" ) a : str = bertabert.config.encoder.vocab_size a : Optional[int] = tokenizer.sep_token_id a : Any = tokenizer.cls_token_id a : str = 128 a : str = datasets.load_dataset("cnn_dailymail" , "3.0.0" , split="train[:1%]" ) a : Union[str, Any] = datasets.load_dataset("cnn_dailymail" , "3.0.0" , split="validation[:1%]" ) a : List[str] = train_dataset.select(range(32 ) ) a : List[Any] = val_dataset.select(range(16 ) ) a : List[str] = 4 def _map_to_encoder_decoder_inputs(lowerCAmelCase__ ): # Tokenizer will automatically set [BOS] <text> [EOS] a : List[str] = tokenizer(batch["article"] , padding="max_length" , truncation=lowerCAmelCase__ , max_length=512 ) a : List[str] = tokenizer(batch["highlights"] , padding="max_length" , truncation=lowerCAmelCase__ , max_length=128 ) a : str = inputs.input_ids a : Tuple = inputs.attention_mask a : Optional[Any] = outputs.input_ids a : List[str] = outputs.input_ids.copy() a : Union[str, Any] = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["labels"] ] a : str = outputs.attention_mask assert all(len(lowerCAmelCase__ ) == 512 for x in inputs.input_ids ) assert all(len(lowerCAmelCase__ ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(lowerCAmelCase__ ): a : List[str] = pred.label_ids a : str = pred.predictions # all unnecessary tokens are removed a : Union[str, Any] = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) a : Optional[Any] = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) a : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(lowerCAmelCase__ ) )] ) / len(lowerCAmelCase__ ) return {"accuracy": accuracy} # map train dataset a : Optional[int] = train_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , remove_columns=["article", "highlights"] , ) train_dataset.set_format( type="torch" , columns=["input_ids", "attention_mask", "decoder_input_ids", "decoder_attention_mask", "labels"] , ) # same for validation dataset a : int = val_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , remove_columns=["article", "highlights"] , ) val_dataset.set_format( type="torch" , columns=["input_ids", "attention_mask", "decoder_input_ids", "decoder_attention_mask", "labels"] , ) a : int = self.get_auto_remove_tmp_dir() a : int = SeqaSeqTrainingArguments( output_dir=lowerCAmelCase__ , per_device_train_batch_size=lowerCAmelCase__ , per_device_eval_batch_size=lowerCAmelCase__ , predict_with_generate=lowerCAmelCase__ , evaluation_strategy="steps" , do_train=lowerCAmelCase__ , do_eval=lowerCAmelCase__ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer a : Optional[Any] = SeqaSeqTrainer( model=lowerCAmelCase__ , args=lowerCAmelCase__ , compute_metrics=_compute_metrics , train_dataset=lowerCAmelCase__ , eval_dataset=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , ) # start training trainer.train()
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a : Optional[int] = { '''configuration_graphormer''': ['''GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GraphormerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Union[str, Any] = [ '''GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GraphormerForGraphClassification''', '''GraphormerModel''', '''GraphormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging a : Tuple = logging.get_logger(__name__) if is_vision_available(): import PIL class __UpperCamelCase ( a__ ): lowerCamelCase : int =["""pixel_values"""] def __init__( self , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = PILImageResampling.BICUBIC , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = 1 / 255 , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = True , **lowerCAmelCase__ , ) -> None: super().__init__(**lowerCAmelCase__ ) a : List[Any] = size if size is not None else {"shortest_edge": 224} a : Optional[Any] = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__ ) a : Any = crop_size if crop_size is not None else {"height": 224, "width": 224} a : str = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__ , param_name="crop_size" ) a : Tuple = do_resize a : Optional[int] = size a : Tuple = resample a : List[str] = do_center_crop a : List[str] = crop_size a : Tuple = do_rescale a : Tuple = rescale_factor a : int = do_normalize a : int = image_mean if image_mean is not None else OPENAI_CLIP_MEAN a : Dict = image_std if image_std is not None else OPENAI_CLIP_STD a : str = do_convert_rgb def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = PILImageResampling.BICUBIC , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: a : List[str] = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__ ) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) a : Optional[Any] = get_resize_output_image_size(lowerCAmelCase__ , size=size["shortest_edge"] , default_to_square=lowerCAmelCase__ ) return resize(lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: a : Tuple = get_size_dict(lowerCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(lowerCAmelCase__ , size=(size["height"], size["width"]) , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Any: return rescale(lowerCAmelCase__ , scale=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> np.ndarray: return normalize(lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = ChannelDimension.FIRST , **lowerCAmelCase__ , ) -> PIL.Image.Image: a : int = do_resize if do_resize is not None else self.do_resize a : Union[str, Any] = size if size is not None else self.size a : int = get_size_dict(lowerCAmelCase__ , param_name="size" , default_to_square=lowerCAmelCase__ ) a : Union[str, Any] = resample if resample is not None else self.resample a : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop a : str = crop_size if crop_size is not None else self.crop_size a : Optional[int] = get_size_dict(lowerCAmelCase__ , param_name="crop_size" , default_to_square=lowerCAmelCase__ ) a : Tuple = do_rescale if do_rescale is not None else self.do_rescale a : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor a : str = do_normalize if do_normalize is not None else self.do_normalize a : Dict = image_mean if image_mean is not None else self.image_mean a : Optional[Any] = image_std if image_std is not None else self.image_std a : str = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb a : str = make_list_of_images(lowerCAmelCase__ ) if not valid_images(lowerCAmelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: a : Any = [convert_to_rgb(lowerCAmelCase__ ) for image in images] # All transformations expect numpy arrays. a : int = [to_numpy_array(lowerCAmelCase__ ) for image in images] if do_resize: a : Dict = [self.resize(image=lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ ) for image in images] if do_center_crop: a : str = [self.center_crop(image=lowerCAmelCase__ , size=lowerCAmelCase__ ) for image in images] if do_rescale: a : List[Any] = [self.rescale(image=lowerCAmelCase__ , scale=lowerCAmelCase__ ) for image in images] if do_normalize: a : Optional[Any] = [self.normalize(image=lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ ) for image in images] a : List[str] = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__ ) for image in images] a : int = {"pixel_values": images} return BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__ )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import math def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if initial_intensity < 0: raise ValueError("The value of intensity cannot be negative" ) # handling of negative values of initial intensity if angle < 0 or angle > 360: raise ValueError("In Malus Law, the angle is in the range 0-360 degrees" ) # handling of values out of allowed range return initial_intensity * (math.cos(math.radians(_lowercase ) ) ** 2) if __name__ == "__main__": import doctest doctest.testmod(name='''malus_law''')
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy a : Union[str, Any] = logging.getLogger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : torch.nn.Module , _lowercase : BnbQuantizationConfig , _lowercase : Union[str, os.PathLike] = None , _lowercase : Optional[Dict[str, Union[int, str, torch.device]]] = None , _lowercase : Optional[List[str]] = None , _lowercase : Optional[Dict[Union[int, str], Union[int, str]]] = None , _lowercase : Optional[Union[str, os.PathLike]] = None , _lowercase : bool = False , ) ->List[Any]: '''simple docstring''' a : Optional[int] = bnb_quantization_config.load_in_abit a : int = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( "You have a version of `bitsandbytes` that is not compatible with 8bit quantization," " make sure you have the latest version of `bitsandbytes` installed." ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( "You have a version of `bitsandbytes` that is not compatible with 4bit quantization," "make sure you have the latest version of `bitsandbytes` installed." ) a : List[str] = [] # custom device map if isinstance(_lowercase , _lowercase ) and len(device_map.keys() ) > 1: a : Any = [key for key, value in device_map.items() if value in ["disk", "cpu"]] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: a : Optional[int] = get_keys_to_not_convert(_lowercase ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(_lowercase ) a : str = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: a : Any = [] a : Optional[int] = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(_lowercase ) # compatibility with peft a : Union[str, Any] = load_in_abit a : List[Any] = load_in_abit a : Optional[Any] = get_parameter_device(_lowercase ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( "It is not recommended to quantize a loaded model. " "The model should be instantiated under the `init_empty_weights` context manager." ) a : Dict = replace_with_bnb_layers(_lowercase , _lowercase , modules_to_not_convert=_lowercase ) # convert param to the right dtype a : Tuple = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: a : int = name.replace(".weight" , "" ).replace(".bias" , "" ) a : Dict = getattr(_lowercase , _lowercase , _lowercase ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(_lowercase ): param.to(_lowercase ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError("No GPU found. A GPU is needed for quantization." ) logger.info( F"""The model device type is {model_device.type}. However, cuda is needed for quantization.""" "We move the model to cuda." ) return model elif weights_location is None: raise RuntimeError( F"""`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} """ ) else: with init_empty_weights(): a : List[Any] = replace_with_bnb_layers( _lowercase , _lowercase , modules_to_not_convert=_lowercase ) a : int = get_quantized_model_device_map( _lowercase , _lowercase , _lowercase , max_memory=_lowercase , no_split_module_classes=_lowercase , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): a : Optional[Any] = True a : List[Any] = any(x in list(device_map.values() ) for x in ["cpu", "disk"] ) load_checkpoint_in_model( _lowercase , _lowercase , _lowercase , dtype=bnb_quantization_config.torch_dtype , offload_folder=_lowercase , offload_state_dict=_lowercase , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(_lowercase , device_map=_lowercase , offload_dir=_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : int , _lowercase : Optional[int]=None , _lowercase : Optional[int]=None , _lowercase : Tuple=None ) ->str: '''simple docstring''' if device_map is None: if torch.cuda.is_available(): a : Dict = {"": torch.cuda.current_device()} else: raise RuntimeError("No GPU found. A GPU is needed for quantization." ) logger.info("The device_map was not initialized." "Setting device_map to `{'':torch.cuda.current_device()}`." ) if isinstance(_lowercase , _lowercase ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( "If passing a string for `device_map`, please choose 'auto', 'balanced', 'balanced_low_0' or " "'sequential'." ) a : List[str] = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) a : List[str] = {} a : Optional[Any] = special_dtypes a : List[Any] = no_split_module_classes a : Optional[Any] = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": a : List[str] = get_balanced_memory( _lowercase , low_zero=(device_map == "balanced_low_0") , max_memory=_lowercase , **_lowercase , ) a : Optional[Any] = max_memory a : str = infer_auto_device_map(_lowercase , **_lowercase ) if isinstance(_lowercase , _lowercase ): # check if don't have any quantized module on the cpu a : Optional[int] = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules a : List[Any] = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( "\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n " ) else: logger.info( "Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit" ) del device_map_without_some_modules return device_map def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Dict , _lowercase : Dict=None , _lowercase : Any=None ) ->Any: '''simple docstring''' if modules_to_not_convert is None: a : Optional[Any] = [] a, a : Optional[int] = _replace_with_bnb_layers( _lowercase , _lowercase , _lowercase , _lowercase ) if not has_been_replaced: logger.warning( "You are loading your model in 8bit or 4bit but no linear modules were found in your model." " this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers." " Please double check your model architecture, or submit an issue on github if you think this is" " a bug." ) return model def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Dict , _lowercase : Any=None , _lowercase : Any=None , ) ->List[Any]: '''simple docstring''' a : Optional[Any] = False for name, module in model.named_children(): if current_key_name is None: a : Tuple = [] current_key_name.append(_lowercase ) if isinstance(_lowercase , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` a : int = ".".join(_lowercase ) a : List[Any] = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: a : Optional[int] = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: a : Union[str, Any] = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=_lowercase , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: a : Optional[int] = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError("load_in_8bit and load_in_4bit can't be both False" ) a : List[Any] = module.weight.data if module.bias is not None: a : Dict = module.bias.data bnb_module.requires_grad_(_lowercase ) setattr(_lowercase , _lowercase , _lowercase ) a : Optional[Any] = True if len(list(module.children() ) ) > 0: a, a : Any = _replace_with_bnb_layers( _lowercase , _lowercase , _lowercase , _lowercase ) a : Dict = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->str: '''simple docstring''' with init_empty_weights(): a : Union[str, Any] = deepcopy(_lowercase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` a : Optional[Any] = find_tied_parameters(_lowercase ) # For compatibility with Accelerate < 0.18 if isinstance(_lowercase , _lowercase ): a : str = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: a : Optional[Any] = sum(_lowercase , [] ) a : List[Any] = len(_lowercase ) > 0 # Check if it is a base model a : Dict = False if hasattr(_lowercase , "base_model_prefix" ): a : Optional[int] = not hasattr(_lowercase , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head a : Union[str, Any] = list(model.named_children() ) a : int = [list_modules[-1][0]] # add last module together with tied weights a : Optional[int] = set(_lowercase ) - set(_lowercase ) a : int = list(set(_lowercase ) ) + list(_lowercase ) # remove ".weight" from the keys a : int = [".weight", ".bias"] a : Optional[int] = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: a : Dict = name.replace(_lowercase , "" ) filtered_module_names.append(_lowercase ) return filtered_module_names def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->Dict: '''simple docstring''' for m in model.modules(): if isinstance(_lowercase , bnb.nn.Linearabit ): return True return False def _SCREAMING_SNAKE_CASE ( _lowercase : nn.Module ) ->List[str]: '''simple docstring''' return next(parameter.parameters() ).device def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int , _lowercase : Tuple , _lowercase : List[str] , _lowercase : List[Any] , _lowercase : Dict , _lowercase : List[str] ) ->int: '''simple docstring''' if fpaa_statistics is None: set_module_tensor_to_device(_lowercase , _lowercase , 0 , dtype=_lowercase , value=_lowercase ) a : int = param_name a : List[str] = model if "." in tensor_name: a : Tuple = tensor_name.split("." ) for split in splits[:-1]: a : int = getattr(_lowercase , _lowercase ) if new_module is None: raise ValueError(F"""{module} has no attribute {split}.""" ) a : List[Any] = new_module a : Dict = splits[-1] # offload weights a : Optional[Any] = False offload_weight(module._parameters[tensor_name] , _lowercase , _lowercase , index=_lowercase ) if hasattr(module._parameters[tensor_name] , "SCB" ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace("weight" , "SCB" ) , _lowercase , index=_lowercase , ) else: offload_weight(_lowercase , _lowercase , _lowercase , index=_lowercase ) offload_weight(_lowercase , param_name.replace("weight" , "SCB" ) , _lowercase , index=_lowercase ) set_module_tensor_to_device(_lowercase , _lowercase , "meta" , dtype=_lowercase , value=torch.empty(*param.size() ) )
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) a : Optional[int] = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongformerConfig''', '''LongformerOnnxConfig''', ], '''tokenization_longformer''': ['''LongformerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''LongformerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = [ '''LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongformerForMaskedLM''', '''LongformerForMultipleChoice''', '''LongformerForQuestionAnswering''', '''LongformerForSequenceClassification''', '''LongformerForTokenClassification''', '''LongformerModel''', '''LongformerPreTrainedModel''', '''LongformerSelfAttention''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : int = [ '''TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLongformerForMaskedLM''', '''TFLongformerForMultipleChoice''', '''TFLongformerForQuestionAnswering''', '''TFLongformerForSequenceClassification''', '''TFLongformerForTokenClassification''', '''TFLongformerModel''', '''TFLongformerPreTrainedModel''', '''TFLongformerSelfAttention''', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys a : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" class __UpperCamelCase : # Public class to implement a graph def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: a : Optional[Any] = row a : Tuple = col a : str = graph def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> bool: return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: # Checking all 8 elements surrounding nth element a : Optional[int] = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order a : List[str] = [-1, 0, 1, -1, 1, -1, 0, 1] a : Dict = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , lowerCAmelCase__ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , lowerCAmelCase__ ) def __a ( self ) -> int: # And finally, count all islands. a : List[str] = [[False for j in range(self.COL )] for i in range(self.ROW )] a : int = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) count += 1 return count
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __UpperCamelCase ( a__ , a__ , a__ , unittest.TestCase ): lowerCamelCase : str =StableUnCLIPPipeline lowerCamelCase : Optional[Any] =TEXT_TO_IMAGE_PARAMS lowerCamelCase : List[str] =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : str =TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase : int =TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false lowerCamelCase : Union[str, Any] =False def __a ( self ) -> Any: a : Optional[int] = 32 a : str = embedder_hidden_size # prior components torch.manual_seed(0 ) a : str = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) a : List[Any] = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowerCAmelCase__ , projection_dim=lowerCAmelCase__ , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) a : Any = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=lowerCAmelCase__ , num_layers=1 , ) torch.manual_seed(0 ) a : Dict = DDPMScheduler( variance_type="fixed_small_log" , prediction_type="sample" , num_train_timesteps=1000 , clip_sample=lowerCAmelCase__ , clip_sample_range=5.0 , beta_schedule="squaredcos_cap_v2" , ) # regular denoising components torch.manual_seed(0 ) a : str = StableUnCLIPImageNormalizer(embedding_dim=lowerCAmelCase__ ) a : Optional[int] = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) a : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) a : Optional[int] = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowerCAmelCase__ , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) a : Dict = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=lowerCAmelCase__ , layers_per_block=1 , upcast_attention=lowerCAmelCase__ , use_linear_projection=lowerCAmelCase__ , ) torch.manual_seed(0 ) a : List[Any] = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.00_085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=lowerCAmelCase__ , steps_offset=1 , ) torch.manual_seed(0 ) a : Union[str, Any] = AutoencoderKL() a : int = { # prior components "prior_tokenizer": prior_tokenizer, "prior_text_encoder": prior_text_encoder, "prior": prior, "prior_scheduler": prior_scheduler, # image noising components "image_normalizer": image_normalizer, "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder, "unet": unet, "scheduler": scheduler, "vae": vae, } return components def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Any: if str(lowerCAmelCase__ ).startswith("mps" ): a : Optional[Any] = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Any = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "prior_num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: a : Optional[int] = torch_device == "cpu" self._test_attention_slicing_forward_pass(test_max_difference=lowerCAmelCase__ ) def __a ( self ) -> Tuple: a : List[str] = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=lowerCAmelCase__ ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> int: a : List[str] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy" ) a : Any = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Tuple = pipe("anime turle" , generator=lowerCAmelCase__ , output_type="np" ) a : Tuple = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() a : List[Any] = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) a : Union[str, Any] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() a : Tuple = pipe( "anime turtle" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="np" , ) a : Optional[Any] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : int , _lowercase : Optional[Any] , _lowercase : Tuple ) ->Optional[Any]: '''simple docstring''' if height >= 1: move_tower(height - 1 , _lowercase , _lowercase , _lowercase ) move_disk(_lowercase , _lowercase ) move_tower(height - 1 , _lowercase , _lowercase , _lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : Any ) ->List[Any]: '''simple docstring''' print("moving disk from" , _lowercase , "to" , _lowercase ) def _SCREAMING_SNAKE_CASE ( ) ->Union[str, Any]: '''simple docstring''' a : Optional[int] = int(input("Height of hanoi: " ).strip() ) move_tower(_lowercase , "A" , "B" , "C" ) if __name__ == "__main__": main()
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : dict ) ->set: '''simple docstring''' a : Tuple = set() # edges = list of graph's edges a : str = get_edges(_lowercase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: a, a : Union[str, Any] = edges.pop() chosen_vertices.add(_lowercase ) chosen_vertices.add(_lowercase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_lowercase ) return chosen_vertices def _SCREAMING_SNAKE_CASE ( _lowercase : dict ) ->set: '''simple docstring''' a : Optional[int] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging a : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : List[str] ) ->Optional[Any]: '''simple docstring''' try: with open(_lowercase , "rb" ) as flax_state_f: a : Union[str, Any] = from_bytes(_lowercase , flax_state_f.read() ) except UnpicklingError as e: try: with open(_lowercase ) as f: if f.read().startswith("version" ): raise OSError( "You seem to have cloned a repository without having git-lfs installed. Please" " install git-lfs and run `git lfs install` followed by `git lfs pull` in the" " folder you cloned." ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(_lowercase , _lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Any , _lowercase : Dict ) ->Tuple: '''simple docstring''' try: import torch # noqa: F401 except ImportError: logger.error( "Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see" " https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation" " instructions." ) raise # check if we have bf16 weights a : str = flatten_dict(jax.tree_util.tree_map(lambda _lowercase : x.dtype == jnp.bfloataa , _lowercase ) ).values() if any(_lowercase ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( "Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` " "before loading those in PyTorch model." ) a : Optional[Any] = jax.tree_util.tree_map( lambda _lowercase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , _lowercase ) a : str = "" a : Union[str, Any] = flatten_dict(_lowercase , sep="." ) a : Optional[int] = pt_model.state_dict() # keep track of unexpected & missing keys a : List[str] = [] a : Dict = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): a : List[Any] = flax_key_tuple.split("." ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: a : Union[str, Any] = flax_key_tuple_array[:-1] + ["weight"] a : int = jnp.transpose(_lowercase , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": a : List[Any] = flax_key_tuple_array[:-1] + ["weight"] a : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": a : Union[str, Any] = flax_key_tuple_array[:-1] + ["weight"] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(_lowercase ): a : Dict = ( flax_key_tuple_string.replace("_0" , ".0" ) .replace("_1" , ".1" ) .replace("_2" , ".2" ) .replace("_3" , ".3" ) .replace("_4" , ".4" ) .replace("_5" , ".5" ) .replace("_6" , ".6" ) .replace("_7" , ".7" ) .replace("_8" , ".8" ) .replace("_9" , ".9" ) ) a : Optional[int] = ".".join(_lowercase ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict a : int = np.asarray(_lowercase ) if not isinstance(_lowercase , np.ndarray ) else flax_tensor a : Optional[Any] = torch.from_numpy(_lowercase ) # remove from missing keys missing_keys.remove(_lowercase ) else: # weight is not expected by PyTorch model unexpected_keys.append(_lowercase ) pt_model.load_state_dict(_lowercase ) # re-transform missing_keys to list a : List[Any] = list(_lowercase ) if len(_lowercase ) > 0: logger.warning( "Some weights of the Flax model were not used when initializing the PyTorch model" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" " (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" " to be exactly identical (e.g. initializing a BertForSequenceClassification model from a" " FlaxBertForSequenceClassification model)." ) if len(_lowercase ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" " use it for predictions and inference." ) return pt_model
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a : Optional[Any] = jnp.ones((batch_size, length) ) / length return scores def __a ( self ) -> List[Any]: a : Optional[Any] = None a : Optional[int] = 20 a : Tuple = self._get_uniform_logits(batch_size=2 , length=lowerCAmelCase__ ) # tweak scores to not be uniform anymore a : List[str] = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch a : Tuple = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax a : Any = jax.nn.softmax(lowerCAmelCase__ , axis=-1 ) a : Tuple = FlaxTemperatureLogitsWarper(temperature=0.5 ) a : Optional[int] = FlaxTemperatureLogitsWarper(temperature=1.3 ) a : Union[str, Any] = jax.nn.softmax(temp_dist_warper_sharper(lowerCAmelCase__ , scores.copy() , cur_len=lowerCAmelCase__ ) , axis=-1 ) a : Tuple = jax.nn.softmax(temp_dist_warper_smoother(lowerCAmelCase__ , scores.copy() , cur_len=lowerCAmelCase__ ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def __a ( self ) -> Any: a : Tuple = None a : Any = 10 a : int = 2 # create ramp distribution a : Dict = np.broadcast_to(np.arange(lowerCAmelCase__ )[None, :] , (batch_size, vocab_size) ).copy() a : Dict = ramp_logits[1:, : vocab_size // 2] + vocab_size a : Dict = FlaxTopKLogitsWarper(3 ) a : int = top_k_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case a : str = 5 a : Optional[int] = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) a : Optional[Any] = np.broadcast_to(np.arange(lowerCAmelCase__ )[None, :] , (batch_size, length) ).copy() a : Union[str, Any] = top_k_warp_safety_check(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def __a ( self ) -> List[str]: a : Optional[Any] = None a : List[str] = 10 a : Tuple = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) a : Optional[Any] = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) a : int = FlaxTopPLogitsWarper(0.8 ) a : Dict = np.exp(top_p_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 a : Any = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # check edge cases with negative and extreme logits a : Dict = np.broadcast_to(np.arange(lowerCAmelCase__ )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme a : Dict = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept a : int = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) a : Optional[Any] = top_p_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def __a ( self ) -> List[Any]: a : Union[str, Any] = 20 a : Optional[int] = 4 a : Dict = 0 a : List[str] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=lowerCAmelCase__ ) # check that min length is applied at length 5 a : Dict = ids_tensor((batch_size, 20) , vocab_size=20 ) a : Optional[int] = 5 a : List[str] = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = min_dist_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float("inf" )] ) # check that min length is not applied anymore at length 15 a : Optional[Any] = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = 15 a : Any = min_dist_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertFalse(jnp.isinf(lowerCAmelCase__ ).any() ) def __a ( self ) -> List[Any]: a : Union[str, Any] = 20 a : Tuple = 4 a : Dict = 0 a : List[str] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=lowerCAmelCase__ ) # check that all scores are -inf except the bos_token_id score a : List[str] = ids_tensor((batch_size, 1) , vocab_size=20 ) a : Optional[int] = 1 a : List[Any] = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : Tuple = logits_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 a : Tuple = 3 a : List[Any] = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : str = logits_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertFalse(jnp.isinf(lowerCAmelCase__ ).any() ) def __a ( self ) -> Union[str, Any]: a : Dict = 20 a : Union[str, Any] = 4 a : int = 0 a : Union[str, Any] = 5 a : int = FlaxForcedEOSTokenLogitsProcessor(max_length=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) # check that all scores are -inf except the eos_token_id when max_length is reached a : List[str] = ids_tensor((batch_size, 4) , vocab_size=20 ) a : List[str] = 4 a : Tuple = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : Union[str, Any] = logits_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached a : List[Any] = 3 a : Any = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : Tuple = logits_processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) self.assertFalse(jnp.isinf(lowerCAmelCase__ ).any() ) def __a ( self ) -> int: a : Optional[Any] = 4 a : Optional[Any] = 10 a : Union[str, Any] = 15 a : Union[str, Any] = 2 a : Any = 1 a : Any = 15 # dummy input_ids and scores a : str = ids_tensor((batch_size, sequence_length) , lowerCAmelCase__ ) a : int = input_ids.copy() a : List[Any] = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[Any] = scores.copy() # instantiate all dist processors a : str = FlaxTemperatureLogitsWarper(temperature=0.5 ) a : Optional[Any] = FlaxTopKLogitsWarper(3 ) a : str = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors a : Any = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=lowerCAmelCase__ ) a : str = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=lowerCAmelCase__ ) a : Optional[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) a : Optional[Any] = 10 # no processor list a : List[Any] = temp_dist_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : List[Any] = top_k_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : List[Any] = top_p_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Optional[Any] = min_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : int = bos_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Optional[int] = eos_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) # with processor list a : List[str] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) a : Optional[int] = processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) # scores should be equal self.assertTrue(jnp.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def __a ( self ) -> Optional[Any]: a : Optional[Any] = 4 a : Any = 10 a : Optional[Any] = 15 a : int = 2 a : List[Any] = 1 a : Dict = 15 # dummy input_ids and scores a : List[str] = ids_tensor((batch_size, sequence_length) , lowerCAmelCase__ ) a : List[str] = input_ids.copy() a : int = self._get_uniform_logits(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = scores.copy() # instantiate all dist processors a : Union[str, Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) a : Union[str, Any] = FlaxTopKLogitsWarper(3 ) a : Tuple = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors a : int = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=lowerCAmelCase__ ) a : Optional[int] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=lowerCAmelCase__ ) a : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) a : List[Any] = 10 # no processor list def run_no_processor_list(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): a : Union[str, Any] = temp_dist_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Optional[int] = top_k_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Optional[Any] = top_p_warp(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Any = min_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : Optional[Any] = bos_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) a : str = eos_dist_proc(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) return scores # with processor list def run_processor_list(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): a : Union[str, Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) a : List[str] = processor(lowerCAmelCase__ , lowerCAmelCase__ , cur_len=lowerCAmelCase__ ) return scores a : Tuple = jax.jit(lowerCAmelCase__ ) a : str = jax.jit(lowerCAmelCase__ ) a : Any = jitted_run_no_processor_list(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) a : Dict = jitted_run_processor_list(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # scores should be equal self.assertTrue(jnp.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=99 , lowerCAmelCase__=32 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=512 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=4 , ) -> Any: a : Dict = parent a : Optional[int] = batch_size a : Tuple = seq_length a : Optional[Any] = is_training a : List[str] = use_attention_mask a : Optional[Any] = use_token_type_ids a : Tuple = use_labels a : Any = vocab_size a : List[Any] = hidden_size a : str = num_hidden_layers a : Tuple = num_attention_heads a : Optional[Any] = intermediate_size a : Union[str, Any] = hidden_act a : List[str] = hidden_dropout_prob a : Optional[Any] = attention_probs_dropout_prob a : Optional[Any] = max_position_embeddings a : Any = type_vocab_size a : List[str] = type_sequence_label_size a : List[Any] = initializer_range a : Optional[int] = num_choices def __a ( self ) -> Optional[Any]: a : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : Optional[Any] = None if self.use_attention_mask: a : int = random_attention_mask([self.batch_size, self.seq_length] ) a : str = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCAmelCase__ , ) return config, input_ids, attention_mask def __a ( self ) -> Optional[Any]: a : Union[str, Any] = self.prepare_config_and_inputs() a, a, a : Optional[Any] = config_and_inputs a : str = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Dict =( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def __a ( self ) -> int: a : str = FlaxDistilBertModelTester(self ) @slow def __a ( self ) -> List[str]: for model_class_name in self.all_model_classes: a : Any = model_class_name.from_pretrained("distilbert-base-uncased" ) a : str = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase__ ) @require_flax class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Optional[int]: a : int = FlaxDistilBertModel.from_pretrained("distilbert-base-uncased" ) a : Dict = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) a : str = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) a : Optional[int] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] a : List[Any] = (1, 11, 768) self.assertEqual(output.shape , lowerCAmelCase__ ) a : List[str] = np.array([[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCAmelCase__ , atol=1E-4 ) )
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" from __future__ import annotations import requests def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->dict: '''simple docstring''' a : List[Any] = F"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(_lowercase ).json() def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 ) ->list[dict]: '''simple docstring''' a : Dict = "https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty" a : str = requests.get(_lowercase ).json()[:max_stories] return [get_hackernews_story(_lowercase ) for story_id in story_ids] def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 ) ->str: '''simple docstring''' a : Dict = hackernews_top_stories(_lowercase ) return "\n".join("* [{title}]({url})".format(**_lowercase ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( ) ->Any: '''simple docstring''' for n in range(1 , 100_0000 ): yield n * (n + 1) // 2 def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->str: '''simple docstring''' a : List[str] = 1 a : Optional[Any] = 2 while i * i <= n: a : Any = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _SCREAMING_SNAKE_CASE ( ) ->str: '''simple docstring''' return next(i for i in triangle_number_generator() if count_divisors(_lowercase ) > 500 ) if __name__ == "__main__": print(solution())
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Optional[int]: a : str = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) a : Dict = torch.tensor([[0, 581, 1_0269, 83, 9_9942, 136, 6_0742, 23, 70, 8_0583, 1_8276, 2]] ) # The dog is cute and lives in the garden house a : List[str] = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim a : Union[str, Any] = torch.tensor( [[-0.0_101, 0.1_218, -0.0_803, 0.0_801, 0.1_327, 0.0_776, -0.1_215, 0.2_383, 0.3_338, 0.3_106, 0.0_300, 0.0_252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): a : Tuple = model(lowerCAmelCase__ )["last_hidden_state"].detach() self.assertEqual(output.shape , lowerCAmelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , lowerCAmelCase__ , atol=1E-3 ) ) @slow def __a ( self ) -> List[str]: a : Optional[int] = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) a : List[str] = torch.tensor([[0, 581, 1_0269, 83, 9_9942, 136, 6_0742, 23, 70, 8_0583, 1_8276, 2]] ) # The dog is cute and lives in the garden house a : List[Any] = torch.Size((1, 12, 1024) ) # batch_size, sequence_length, embedding_vector_dim a : int = torch.tensor( [[-0.0_699, -0.0_318, 0.0_705, -0.1_241, 0.0_999, -0.0_520, 0.1_004, -0.1_838, -0.4_704, 0.1_437, 0.0_821, 0.0_126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): a : Tuple = model(lowerCAmelCase__ )["last_hidden_state"].detach() self.assertEqual(output.shape , lowerCAmelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , lowerCAmelCase__ , atol=1E-3 ) )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import colorsys from PIL import Image # type: ignore def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float , _lowercase : int ) ->float: '''simple docstring''' a : Tuple = x a : Union[str, Any] = y for step in range(_lowercase ): # noqa: B007 a : Any = a * a - b * b + x a : str = 2 * a * b + y a : Dict = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _SCREAMING_SNAKE_CASE ( _lowercase : float ) ->tuple: '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def _SCREAMING_SNAKE_CASE ( _lowercase : float ) ->tuple: '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255 ) for i in colorsys.hsv_to_rgb(_lowercase , 1 , 1 ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : int = 800 , _lowercase : int = 600 , _lowercase : float = -0.6 , _lowercase : float = 0 , _lowercase : float = 3.2 , _lowercase : int = 50 , _lowercase : bool = True , ) ->Image.Image: '''simple docstring''' a : List[str] = Image.new("RGB" , (image_width, image_height) ) a : Union[str, Any] = img.load() # loop through the image-coordinates for image_x in range(_lowercase ): for image_y in range(_lowercase ): # determine the figure-coordinates based on the image-coordinates a : List[Any] = figure_width / image_width * image_height a : Any = figure_center_x + (image_x / image_width - 0.5) * figure_width a : List[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height a : Union[str, Any] = get_distance(_lowercase , _lowercase , _lowercase ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: a : List[Any] = get_color_coded_rgb(_lowercase ) else: a : List[Any] = get_black_and_white_rgb(_lowercase ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure a : Union[str, Any] = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def _SCREAMING_SNAKE_CASE ( _lowercase : jnp.ndarray , _lowercase : int , _lowercase : float = 1 , _lowercase : float = 1 , _lowercase : float = 1.0E4 , _lowercase : bool = False , _lowercase : float = 1.0 , ) ->jnp.ndarray: '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, F"""Embedding dimension {embedding_dim} should be even""" a : Union[str, Any] = float(embedding_dim // 2 ) a : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) a : List[str] = min_timescale * jnp.exp(jnp.arange(_lowercase , dtype=jnp.floataa ) * -log_timescale_increment ) a : int = jnp.expand_dims(_lowercase , 1 ) * jnp.expand_dims(_lowercase , 0 ) # scale embeddings a : int = scale * emb if flip_sin_to_cos: a : Tuple = jnp.concatenate([jnp.cos(_lowercase ), jnp.sin(_lowercase )] , axis=1 ) else: a : Dict = jnp.concatenate([jnp.sin(_lowercase ), jnp.cos(_lowercase )] , axis=1 ) a : Dict = jnp.reshape(_lowercase , [jnp.shape(_lowercase )[0], embedding_dim] ) return signal class __UpperCamelCase ( nn.Module ): lowerCamelCase : int =32 lowerCamelCase : jnp.dtype =jnp.floataa @nn.compact def __call__( self , lowerCAmelCase__ ) -> Any: a : Optional[int] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="linear_1" )(lowerCAmelCase__ ) a : int = nn.silu(lowerCAmelCase__ ) a : Optional[int] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="linear_2" )(lowerCAmelCase__ ) return temb class __UpperCamelCase ( nn.Module ): lowerCamelCase : int =32 lowerCamelCase : bool =False lowerCamelCase : float =1 @nn.compact def __call__( self , lowerCAmelCase__ ) -> Union[str, Any]: return get_sinusoidal_embeddings( lowerCAmelCase__ , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def _SCREAMING_SNAKE_CASE ( ) ->Generator[int, None, None]: '''simple docstring''' a : dict[int, int] = {} a : Dict = 2 while True: a : Optional[Any] = factor_map.pop(_lowercase , _lowercase ) if factor: a : Tuple = factor + prime while x in factor_map: x += factor a : Optional[Any] = factor else: a : int = prime yield prime prime += 1 def _SCREAMING_SNAKE_CASE ( _lowercase : float = 1E1_0 ) ->int: '''simple docstring''' a : Tuple = sieve() a : str = 1 while True: a : Optional[int] = next(_lowercase ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(_lowercase ) n += 2 if __name__ == "__main__": print(solution())
31
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : Dict =IFPipeline lowerCamelCase : int =TEXT_TO_IMAGE_PARAMS - {"""width""", """height""", """latents"""} lowerCamelCase : int =TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : int =PipelineTesterMixin.required_optional_params - {"""latents"""} def __a ( self ) -> List[str]: return self._get_dummy_components() def __a ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ) -> Dict: if str(lowerCAmelCase__ ).startswith("mps" ): a : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: a : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) a : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self ) -> Union[str, Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self ) -> Any: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ) -> Union[str, Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ) -> Optional[int]: self._test_save_load_local() def __a ( self ) -> Tuple: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ) -> Tuple: # if a : Tuple = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) a : str = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) a, a : List[str] = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() a : Optional[int] = None a : Optional[int] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img a : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) a : List[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting a : Union[str, Any] = IFInpaintingPipeline(**pipe_a.components ) a : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: # pipeline 1 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (64, 64, 3) a : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Union[str, Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: # pipeline 1 _start_torch_memory_measurement() a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : Tuple = output.images[0] assert image.shape == (64, 64, 3) a : int = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : int = output.images[0] assert image.shape == (256, 256, 3) a : str = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: # pipeline 1 _start_torch_memory_measurement() a : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : List[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) a : List[str] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , num_inference_steps=2 , generator=lowerCAmelCase__ , output_type="np" , ) a : List[Any] = output.images[0] assert image.shape == (64, 64, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 a : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) # pipeline 2 _start_torch_memory_measurement() a : str = torch.Generator(device="cpu" ).manual_seed(0 ) a : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : int = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) a : Dict = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(lowerCAmelCase__ ) a : Optional[int] = pipe_a( prompt_embeds=lowerCAmelCase__ , negative_prompt_embeds=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , original_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) a : List[str] = output.images[0] assert image.shape == (256, 256, 3) a : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 a : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( ) ->List[str]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
31
1
"""simple docstring""" import cva import numpy as np class __UpperCamelCase : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: if k in (0.04, 0.06): a : int = k a : Dict = window_size else: raise ValueError("invalid k value" ) def __str__( self ) -> str: return str(self.k ) def __a ( self , lowerCAmelCase__ ) -> tuple[cva.Mat, list[list[int]]]: a : Optional[int] = cva.imread(lowerCAmelCase__ , 0 ) a, a : List[Any] = img.shape a : list[list[int]] = [] a : List[Any] = img.copy() a : Union[str, Any] = cva.cvtColor(lowerCAmelCase__ , cva.COLOR_GRAY2RGB ) a, a : Optional[int] = np.gradient(lowerCAmelCase__ ) a : int = dx**2 a : Any = dy**2 a : Any = dx * dy a : Dict = 0.04 a : List[Any] = self.window_size // 2 for y in range(lowerCAmelCase__ , h - offset ): for x in range(lowerCAmelCase__ , w - offset ): a : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() a : List[str] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() a : Any = (wxx * wyy) - (wxy**2) a : int = wxx + wyy a : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": a : int = HarrisCorner(0.04, 3) a , a : str = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
31
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
1
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Tuple: a : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Optional[Any]: a, a, a, a : List[str] = hidden_states.shape a : List[Any] = jax.image.resize( lowerCAmelCase__ , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) a : List[str] = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> Dict: a : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ ) -> Tuple: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a : Tuple = self.conv(lowerCAmelCase__ ) return hidden_states class __UpperCamelCase ( nn.Module ): lowerCamelCase : int lowerCamelCase : int =None lowerCamelCase : float =0.0 lowerCamelCase : bool =None lowerCamelCase : jnp.dtype =jnp.floataa def __a ( self ) -> int: a : Dict = self.in_channels if self.out_channels is None else self.out_channels a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : List[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : List[Any] = nn.Dense(lowerCAmelCase__ , dtype=self.dtype ) a : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) a : Optional[int] = nn.Dropout(self.dropout_prob ) a : Dict = nn.Conv( lowerCAmelCase__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a : Union[str, Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a : List[str] = None if use_nin_shortcut: a : Optional[Any] = nn.Conv( lowerCAmelCase__ , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True ) -> str: a : int = hidden_states a : Tuple = self.norma(lowerCAmelCase__ ) a : Any = nn.swish(lowerCAmelCase__ ) a : int = self.conva(lowerCAmelCase__ ) a : int = self.time_emb_proj(nn.swish(lowerCAmelCase__ ) ) a : Tuple = jnp.expand_dims(jnp.expand_dims(lowerCAmelCase__ , 1 ) , 1 ) a : Dict = hidden_states + temb a : str = self.norma(lowerCAmelCase__ ) a : List[Any] = nn.swish(lowerCAmelCase__ ) a : List[str] = self.dropout(lowerCAmelCase__ , lowerCAmelCase__ ) a : List[str] = self.conva(lowerCAmelCase__ ) if self.conv_shortcut is not None: a : Tuple = self.conv_shortcut(lowerCAmelCase__ ) return hidden_states + residual
31
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor a : Union[str, Any] = logging.get_logger(__name__) class __UpperCamelCase ( a__ ): def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> None: warnings.warn( "The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use MobileViTImageProcessor instead." , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
31
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _SCREAMING_SNAKE_CASE ( _lowercase : str=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : Dict = subparsers.add_parser("test" ) else: a : Tuple = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=_lowercase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->str: '''simple docstring''' a : List[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: a : int = script_name else: a : int = F"""--config_file={args.config_file} {script_name}""" a : Optional[int] = ["accelerate-launch"] + test_args.split() a : Optional[int] = execute_subprocess_async(_lowercase , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : Any = test_command_parser() a : Union[str, Any] = parser.parse_args() test_command(_lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
"""simple docstring""" a : str = 8.314_4598 def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a : Any = 300 a : Dict = 28 a : Dict = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
31
1
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig a : Optional[Any] = logging.get_logger(__name__) a : List[str] = '''T5Config''' class __UpperCamelCase ( a__ ): lowerCamelCase : int ="""mt5""" lowerCamelCase : List[Any] =MTaConfig class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] ="""mt5""" lowerCamelCase : List[Any] =MTaConfig class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[int] ="""mt5""" lowerCamelCase : Optional[Any] =MTaConfig
31
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ ) -> Optional[int]: a : str = 3 a : str = 250 a : List[Any] = ids_tensor((batch_size, length) , lowerCAmelCase__ ) a : Optional[Any] = torch.ones((batch_size, length) , device=lowerCAmelCase__ , dtype=torch.float ) / length return input_ids, scores def __a ( self ) -> List[Any]: a, a : str = self._get_tensors(5 ) a : Any = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : str = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = MaxLengthCriteria(max_length=10 ) a, a : int = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : Union[str, Any] = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : Tuple = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) a, a : str = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a, a : int = self._get_tensors(10 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : List[Any] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 10 ) def __a ( self ) -> str: a, a : Tuple = self._get_tensors(5 ) a : str = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) a : Optional[int] = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase__ , lowerCAmelCase__ ) ) def __a ( self ) -> str: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 10 ) with self.assertWarns(lowerCAmelCase__ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10 )] ) , 11 ) a : Optional[int] = validate_stopping_criteria(StoppingCriteriaList() , 11 ) self.assertEqual(len(lowerCAmelCase__ ) , 1 )
31
1
"""simple docstring""" import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( SeqaSeqDataset, calculate_bleu, calculate_rouge, chunks, lmap, load_json, parse_numeric_n_bool_cl_kwargs, save_json, use_task_specific_params, write_txt_file, ) a : Any = getLogger(__name__) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : str , _lowercase : str , _lowercase : int = 8 , _lowercase : int = 1024 , _lowercase : List[str]="val" , _lowercase : Union[str, Any]=None , _lowercase : str=False , _lowercase : Optional[Any]="summarization" , _lowercase : int=None , _lowercase : Any=1 , _lowercase : Dict = None , _lowercase : int="" , **_lowercase : str , ) ->Dict: '''simple docstring''' a : Optional[Any] = str(_lowercase ) assert local_rank is not None torch.distributed.init_process_group(backend="nccl" , rank=_lowercase ) a : str = Path(_lowercase ) a : int = save_dir.joinpath(F"""rank_{local_rank}_output.json""" ) torch.cuda.set_device(_lowercase ) a : Optional[Any] = AutoModelForSeqaSeqLM.from_pretrained(_lowercase ).cuda() if fpaa: a : Any = model.half() # determine if we need to increase num_beams use_task_specific_params(_lowercase , _lowercase ) # update config with task specific params a : Optional[int] = generate_kwargs.pop("num_beams" , model.config.num_beams ) # AttributeError risk? if num_return_sequences > num_beams: a : Union[str, Any] = num_return_sequences a : List[Any] = AutoTokenizer.from_pretrained(_lowercase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. if max_source_length is None: a : Union[str, Any] = tokenizer.model_max_length if prefix is None: a : Optional[Any] = prefix or getattr(model.config , "prefix" , "" ) or "" a : List[str] = SeqaSeqDataset( _lowercase , _lowercase , _lowercase , max_target_length=1024 , type_path=_lowercase , n_obs=_lowercase , prefix=_lowercase , **_lowercase , ) # I set shuffle=True for a more accurate progress bar. # If all the longest samples are first, the prog bar estimate is too high at the beginning. a : Dict = ds.make_sortish_sampler(_lowercase , distributed=_lowercase , add_extra_examples=_lowercase , shuffle=_lowercase ) a : Optional[int] = DataLoader(_lowercase , sampler=_lowercase , batch_size=_lowercase , collate_fn=ds.collate_fn ) a : List[Any] = [] for batch in tqdm(_lowercase ): a : List[str] = model.generate( input_ids=batch["input_ids"].to(model.device ) , attention_mask=batch["attention_mask"].to(model.device ) , num_return_sequences=_lowercase , num_beams=_lowercase , **_lowercase , ) a : Optional[Any] = tokenizer.batch_decode(_lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase ) a : Tuple = batch["ids"] if num_return_sequences > 1: a : Union[str, Any] = chunks(_lowercase , _lowercase ) # batch size chunks, each of size num_return_seq for i, pred in enumerate(_lowercase ): results.append({"pred": pred, "id": ids[i].item()} ) save_json(_lowercase , _lowercase ) return results, sampler.num_replicas def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[str] = argparse.ArgumentParser( epilog="Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate" ) parser.add_argument("--data_dir" , type=_lowercase , help="like cnn_dm/test.source" ) parser.add_argument( "--model_name" , type=_lowercase , help="like facebook/bart-large-cnn,t5-base, etc." , default="sshleifer/distilbart-xsum-12-3" , ) parser.add_argument("--save_dir" , type=_lowercase , help="where to save" , default="tmp_gen" ) parser.add_argument("--max_source_length" , type=_lowercase , default=_lowercase ) parser.add_argument( "--type_path" , type=_lowercase , default="test" , help="which subset to evaluate typically train/val/test" ) parser.add_argument("--task" , type=_lowercase , default="summarization" , help="used for task_specific_params + metrics" ) parser.add_argument("--bs" , type=_lowercase , default=8 , required=_lowercase , help="batch size" ) parser.add_argument( "--local_rank" , type=_lowercase , default=-1 , required=_lowercase , help="should be passed by distributed.launch" ) parser.add_argument( "--n_obs" , type=_lowercase , default=_lowercase , required=_lowercase , help="How many observations. Defaults to all." ) parser.add_argument( "--num_return_sequences" , type=_lowercase , default=1 , required=_lowercase , help="How many sequences to return" ) parser.add_argument( "--sync_timeout" , type=_lowercase , default=600 , required=_lowercase , help="How long should master process wait for other processes to finish." , ) parser.add_argument("--src_lang" , type=_lowercase , default=_lowercase , required=_lowercase ) parser.add_argument("--tgt_lang" , type=_lowercase , default=_lowercase , required=_lowercase ) parser.add_argument( "--prefix" , type=_lowercase , required=_lowercase , default=_lowercase , help="will be added to the begininng of src examples" ) parser.add_argument("--fp16" , action="store_true" ) parser.add_argument("--debug" , action="store_true" ) a : str = time.time() a, a : Any = parser.parse_known_args() a : Tuple = parse_numeric_n_bool_cl_kwargs(_lowercase ) if generate_kwargs and args.local_rank <= 0: print(F"""parsed the following generate kwargs: {generate_kwargs}""" ) a : str = Path(args.save_dir + "_tmp" ) Path(_lowercase ).mkdir(exist_ok=_lowercase ) # this handles locking. a : List[Any] = list(json_save_dir.glob("rank_*.json" ) ) if intermediate_files: raise ValueError(F"""Found files at {json_save_dir} please move or remove them.""" ) # In theory, a node could finish and save before another node hits this. If this happens, we can address later. a : str = {} if args.src_lang is not None: a : List[Any] = args.src_lang if args.tgt_lang is not None: a : Dict = args.tgt_lang Path(args.save_dir ).mkdir(exist_ok=_lowercase ) a, a : Any = eval_data_dir( args.data_dir , _lowercase , args.model_name , type_path=args.type_path , bs=args.bs , fpaa=args.fpaa , task=args.task , local_rank=args.local_rank , n_obs=args.n_obs , max_source_length=args.max_source_length , num_return_sequences=args.num_return_sequences , prefix=args.prefix , dataset_kwargs=_lowercase , **_lowercase , ) if args.local_rank <= 0: a : List[Any] = Path(args.save_dir ) save_dir.mkdir(exist_ok=_lowercase ) a : Union[str, Any] = gather_results_from_each_node(_lowercase , _lowercase , args.sync_timeout ) a : str = combine_partial_results(_lowercase ) if args.num_return_sequences > 1: a : int = save_dir.joinpath("pseudolabel_results.json" ) print(F"""Saving aggregated results at {save_path}, intermediate in {json_save_dir}/""" ) save_json(_lowercase , _lowercase ) return a : Tuple = Path(args.data_dir ).joinpath(args.type_path + ".target" ) with open(_lowercase ) as f: a : str = [x.rstrip() for x in f.readlines()][: len(_lowercase )] # Calculate metrics, save metrics, and save _generations.txt a : Any = "translation" in args.task a : Tuple = calculate_bleu if calc_bleu else calculate_rouge a : Tuple = "bleu" if calc_bleu else "rouge" a : Dict = score_fn(_lowercase , _lowercase ) a : Dict = len(_lowercase ) a : Optional[int] = time.time() - start_time a : Tuple = round(runtime / metrics["n_obs"] , 4 ) a : Optional[Any] = num_replicas # TODO(@stas00): add whatever metadata to metrics a : Any = save_dir.joinpath(F"""{args.type_path}_{metric_name}.json""" ) save_json(_lowercase , _lowercase , indent=_lowercase ) print(_lowercase ) write_txt_file(_lowercase , save_dir.joinpath(F"""{args.type_path}_generations.txt""" ) ) if args.debug: write_txt_file(_lowercase , save_dir.joinpath(F"""{args.type_path}.target""" ) ) else: shutil.rmtree(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->List: '''simple docstring''' a : List[str] = [] for partial_result in partial_results: records.extend(_lowercase ) a : str = sorted(_lowercase , key=lambda _lowercase : x["id"] ) a : List[Any] = [x["pred"] for x in records] return preds def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[int] , _lowercase : str ) ->List[Dict[str, List]]: '''simple docstring''' a : Optional[Any] = time.time() logger.info("waiting for all nodes to finish" ) a : str = None while (time.time() - start_wait) < timeout: a : int = list(save_dir.glob("rank_*.json" ) ) if len(_lowercase ) < num_replicas: continue try: # make sure all json files are fully saved a : Tuple = lmap(_lowercase , _lowercase ) return json_data except JSONDecodeError: continue else: raise TimeoutError("Rank 0 gave up on waiting for other processes" ) # Unreachable if __name__ == "__main__": # Usage for MT: run_generate()
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200 ) ->int: '''simple docstring''' a : Dict = [1, 2, 5, 10, 20, 50, 100, 200] a : Optional[Any] = [0] * (pence + 1) a : List[Any] = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
31
1
"""simple docstring""" import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class __UpperCamelCase ( unittest.TestCase ): def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: return f"""gaussian_noise_s={seed}_shape={'_'.join([str(lowerCAmelCase__ ) for s in shape] )}.npy""" def __a ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() def __a ( self , lowerCAmelCase__=0 , lowerCAmelCase__=(4, 4, 64, 64) , lowerCAmelCase__=False ) -> Tuple: a : List[Any] = jnp.bfloataa if fpaa else jnp.floataa a : Tuple = jnp.array(load_hf_numpy(self.get_file_format(lowerCAmelCase__ , lowerCAmelCase__ ) ) , dtype=lowerCAmelCase__ ) return image def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__="CompVis/stable-diffusion-v1-4" ) -> Union[str, Any]: a : Dict = jnp.bfloataa if fpaa else jnp.floataa a : Tuple = "bf16" if fpaa else None a, a : Optional[int] = FlaxUNetaDConditionModel.from_pretrained( lowerCAmelCase__ , subfolder="unet" , dtype=lowerCAmelCase__ , revision=lowerCAmelCase__ ) return model, params def __a ( self , lowerCAmelCase__=0 , lowerCAmelCase__=(4, 77, 768) , lowerCAmelCase__=False ) -> List[Any]: a : int = jnp.bfloataa if fpaa else jnp.floataa a : Optional[Any] = jnp.array(load_hf_numpy(self.get_file_format(lowerCAmelCase__ , lowerCAmelCase__ ) ) , dtype=lowerCAmelCase__ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_323, -0.1_304, 0.0_813, -0.3_093, -0.0_919, -0.1_571, -0.1_125, -0.5_806]], [17, 0.55, [-0.0_831, -0.2_443, 0.0_901, -0.0_919, 0.3_396, 0.0_103, -0.3_743, 0.0_701]], [8, 0.89, [-0.4_863, 0.0_859, 0.0_875, -0.1_658, 0.9_199, -0.0_114, 0.4_839, 0.4_639]], [3, 1000, [-0.5_649, 0.2_402, -0.5_518, 0.1_248, 1.1_328, -0.2_443, -0.0_325, -1.0_078]], # fmt: on ] ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: a, a : List[Any] = self.get_unet_model(model_id="CompVis/stable-diffusion-v1-4" , fpaa=lowerCAmelCase__ ) a : List[Any] = self.get_latents(lowerCAmelCase__ , fpaa=lowerCAmelCase__ ) a : str = self.get_encoder_hidden_states(lowerCAmelCase__ , fpaa=lowerCAmelCase__ ) a : int = model.apply( {"params": params} , lowerCAmelCase__ , jnp.array(lowerCAmelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCAmelCase__ , ).sample assert sample.shape == latents.shape a : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) a : str = jnp.array(lowerCAmelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_514, 0.0_807, 0.1_624, 0.1_016, -0.1_896, 0.0_263, 0.0_677, 0.2_310]], [17, 0.55, [0.1_164, -0.0_216, 0.0_170, 0.1_589, -0.3_120, 0.1_005, -0.0_581, -0.1_458]], [8, 0.89, [-0.1_758, -0.0_169, 0.1_004, -0.1_411, 0.1_312, 0.1_103, -0.1_996, 0.2_139]], [3, 1000, [0.1_214, 0.0_352, -0.0_731, -0.1_562, -0.0_994, -0.0_906, -0.2_340, -0.0_539]], # fmt: on ] ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: a, a : int = self.get_unet_model(model_id="stabilityai/stable-diffusion-2" , fpaa=lowerCAmelCase__ ) a : Dict = self.get_latents(lowerCAmelCase__ , shape=(4, 4, 96, 96) , fpaa=lowerCAmelCase__ ) a : List[Any] = self.get_encoder_hidden_states(lowerCAmelCase__ , shape=(4, 77, 1024) , fpaa=lowerCAmelCase__ ) a : Optional[int] = model.apply( {"params": params} , lowerCAmelCase__ , jnp.array(lowerCAmelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCAmelCase__ , ).sample assert sample.shape == latents.shape a : int = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) a : Optional[Any] = jnp.array(lowerCAmelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-2 )
31
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 22 ) ->int: '''simple docstring''' a : Optional[Any] = range(1 , _lowercase ) a : Tuple = range(1 , _lowercase ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F'''{solution(10, 22) = }''')
31
"""simple docstring""" import qiskit def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->qiskit.result.counts.Counts: '''simple docstring''' a : Union[str, Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a : Optional[Any] = qiskit.QuantumCircuit(_lowercase , _lowercase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator a : Optional[int] = qiskit.execute(_lowercase , _lowercase , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowercase ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
31
1
"""simple docstring""" from __future__ import annotations from typing import TypedDict class __UpperCamelCase ( a__ ): lowerCamelCase : str lowerCamelCase : int def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->list[str]: '''simple docstring''' if not isinstance(_lowercase , _lowercase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(_lowercase ) )] def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->BWTTransformDict: '''simple docstring''' if not isinstance(_lowercase , _lowercase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) a : str = all_rotations(_lowercase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation a : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(_lowercase ), } return response def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : int ) ->str: '''simple docstring''' if not isinstance(_lowercase , _lowercase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: a : Tuple = int(_lowercase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(_lowercase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) a : str = [""] * len(_lowercase ) for _ in range(len(_lowercase ) ): for i in range(len(_lowercase ) ): a : int = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": a : Optional[int] = '''Provide a string that I will generate its BWT transform: ''' a : Optional[int] = input(entry_msg).strip() a : List[Any] = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result["bwt_string"]}\'''' ) a : Any = reverse_bwt(result['''bwt_string'''], result['''idx_original_string''']) print( F'''Reversing Burrows Wheeler transform for entry \'{result["bwt_string"]}\' ''' F'''we get original string \'{original_string}\'''' )
31
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any] ) ->Dict: '''simple docstring''' a : List[str] = 0 if start < end: a : Tuple = randint(_lowercase , _lowercase ) a : List[str] = a[end] a : str = a[pivot] a : Optional[int] = temp a, a : Dict = _in_place_partition(_lowercase , _lowercase , _lowercase ) count += _in_place_quick_sort(_lowercase , _lowercase , p - 1 ) count += _in_place_quick_sort(_lowercase , p + 1 , _lowercase ) return count def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : List[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 0 a : List[Any] = randint(_lowercase , _lowercase ) a : int = a[end] a : List[str] = a[pivot] a : Tuple = temp a : Union[str, Any] = start - 1 for index in range(_lowercase , _lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value a : List[str] = new_pivot_index + 1 a : Optional[int] = a[new_pivot_index] a : Union[str, Any] = a[index] a : List[Any] = temp a : Tuple = a[new_pivot_index + 1] a : str = a[end] a : Dict = temp return new_pivot_index + 1, count a : int = TemporaryFile() a : Tuple = 100 # 1000 elements are to be sorted a , a : int = 0, 1 # mean and standard deviation a : List[Any] = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array a : int = np.load(outfile) a : Tuple = len(M) - 1 a : Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
31
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Dict: a : Tuple = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) a : Any = { "input_ids": tf.convert_to_tensor([[0, 2646, 1_0269, 83, 9_9942, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } a : str = model(lowerCAmelCase__ )["last_hidden_state"] a : Optional[int] = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , lowerCAmelCase__ ) # compare the actual values for a slice. a : List[str] = tf.convert_to_tensor( [ [ [0.0_681_762, 0.10_894_451, 0.06_772_504], [-0.06_423_668, 0.02_366_615, 0.04_329_344], [-0.06_057_295, 0.09_974_135, -0.00_070_584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
31
"""simple docstring""" import baseaa def _SCREAMING_SNAKE_CASE ( _lowercase : str ) ->bytes: '''simple docstring''' return baseaa.aaaencode(string.encode("utf-8" ) ) def _SCREAMING_SNAKE_CASE ( _lowercase : bytes ) ->str: '''simple docstring''' return baseaa.aaadecode(_lowercase ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
31
1
"""simple docstring""" import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class __UpperCamelCase ( a__ , a__ , unittest.TestCase ): lowerCamelCase : List[str] =VQModel lowerCamelCase : List[str] ="""sample""" @property def __a ( self , lowerCAmelCase__=(32, 32) ) -> Any: a : Dict = 4 a : str = 3 a : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCAmelCase__ ) return {"sample": image} @property def __a ( self ) -> Union[str, Any]: return (3, 32, 32) @property def __a ( self ) -> int: return (3, 32, 32) def __a ( self ) -> int: a : str = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 3, } a : Union[str, Any] = self.dummy_input return init_dict, inputs_dict def __a ( self ) -> List[Any]: pass def __a ( self ) -> Union[str, Any]: pass def __a ( self ) -> int: a, a : Optional[int] = VQModel.from_pretrained("fusing/vqgan-dummy" , output_loading_info=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCAmelCase__ ) a : Optional[int] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __a ( self ) -> int: a : List[str] = VQModel.from_pretrained("fusing/vqgan-dummy" ) model.to(lowerCAmelCase__ ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) a : int = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) a : Union[str, Any] = image.to(lowerCAmelCase__ ) with torch.no_grad(): a : Dict = model(lowerCAmelCase__ ).sample a : Tuple = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off a : Tuple = torch.tensor([-0.0_153, -0.4_044, -0.1_880, -0.5_161, -0.2_418, -0.4_072, -0.1_612, -0.0_633, -0.0_143] ) # fmt: on self.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: a : Tuple = None a : Any = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} a : str = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } a : str = { '''albert-base-v1''': 512, '''albert-large-v1''': 512, '''albert-xlarge-v1''': 512, '''albert-xxlarge-v1''': 512, '''albert-base-v2''': 512, '''albert-large-v2''': 512, '''albert-xlarge-v2''': 512, '''albert-xxlarge-v2''': 512, } a : Union[str, Any] = '''▁''' class __UpperCamelCase ( a__ ): lowerCamelCase : Union[str, Any] =VOCAB_FILES_NAMES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] =AlbertTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , **lowerCAmelCase__ , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. a : Optional[int] = ( AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ , normalized=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token ) super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : Dict = do_lower_case a : Any = remove_space a : Optional[Any] = keep_accents a : List[str] = vocab_file a : Optional[Any] = False if not self.vocab_file else True def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : Optional[Any] = [self.sep_token_id] a : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return a : Dict = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): @slow def __a ( self ) -> Any: a : int = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) a : Any = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 2_5543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" a : List[Any] = model(lowerCAmelCase__ )["last_hidden_state"] a : Optional[Any] = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , lowerCAmelCase__ ) # compare the actual values for a slice. a : str = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
31
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch a : List[Any] = random.Random() def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : int=1.0 , _lowercase : Optional[int]=None , _lowercase : Union[str, Any]=None ) ->Optional[Any]: '''simple docstring''' if rng is None: a : Tuple = global_rng a : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=400 , lowerCAmelCase__=2000 , lowerCAmelCase__=1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=1_6000 , lowerCAmelCase__=True , lowerCAmelCase__=80 , lowerCAmelCase__=16 , lowerCAmelCase__=64 , lowerCAmelCase__="hann_window" , lowerCAmelCase__=80 , lowerCAmelCase__=7600 , lowerCAmelCase__=1E-10 , lowerCAmelCase__=True , ) -> Optional[Any]: a : int = parent a : Tuple = batch_size a : Dict = min_seq_length a : Any = max_seq_length a : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a : Union[str, Any] = feature_size a : Tuple = padding_value a : str = sampling_rate a : Dict = do_normalize a : str = num_mel_bins a : List[str] = hop_length a : str = win_length a : Optional[Any] = win_function a : List[str] = fmin a : Any = fmax a : Optional[int] = mel_floor a : Tuple = return_attention_mask def __a ( self ) -> Optional[Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Tuple: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: a : Union[str, Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a : str = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Any = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs def __a ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: if equal_length: a : str = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a : Any = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a : Optional[int] = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch class __UpperCamelCase ( a__ , unittest.TestCase ): lowerCamelCase : Tuple =SpeechTaFeatureExtractor def __a ( self ) -> Union[str, Any]: a : Tuple = SpeechTaFeatureExtractionTester(self ) def __a ( self , lowerCAmelCase__ ) -> Union[str, Any]: self.assertTrue(np.all(np.mean(lowerCAmelCase__ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCAmelCase__ , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> Union[str, Any]: # Tests that all call wrap to encode_plus and batch_encode_plus a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Any = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input a : Optional[int] = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values a : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values a : int = feat_extract(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> Optional[Any]: a : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = ["longest", "max_length", "do_not_pad"] a : Tuple = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : Dict = feat_extract(lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> str: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : List[str] = range(800 , 1400 , 200 ) a : List[str] = [floats_list((1, x) )[0] for x in lengths] a : Any = ["longest", "max_length", "do_not_pad"] a : Any = [None, 1600, None] for max_length, padding in zip(lowerCAmelCase__ , lowerCAmelCase__ ): a : List[Any] = feat_extract(lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding=lowerCAmelCase__ ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Dict: a : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Union[str, Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="max_length" , return_tensors="np" ) a : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: a : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : List[Any] = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=1000 , padding="longest" , return_tensors="np" ) a : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) a : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : int = feat_extract( lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=2000 , padding="longest" , return_tensors="np" ) a : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> List[str]: a : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a : Any = np.random.rand(100 ).astype(np.floataa ) a : Optional[Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a : str = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a : List[str] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus a : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] a : Tuple = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test feature size a : Union[str, Any] = feature_extractor(audio_target=lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a : Dict = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values a : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test batched a : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : Any = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a : Optional[Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] a : List[Any] = np.asarray(lowerCAmelCase__ ) a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values a : str = feature_extractor(lowerCAmelCase__ , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def __a ( self ) -> str: a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Any = self.feature_extraction_class(**self.feat_extract_dict ) a : Union[str, Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) for x, y in zip(lowerCAmelCase__ , processed_features[input_name] ) ) ) a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : Dict = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Tuple: a : Tuple = self.feat_extract_tester.prepare_inputs_for_target(equal_length=lowerCAmelCase__ ) a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) a : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: a : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) a : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target() a : Optional[Any] = feat_extract.model_input_names[0] a : List[str] = BatchFeature({input_name: speech_inputs} ) a : Tuple = feat_extract.num_mel_bins # hack! a : List[Any] = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" )[input_name] a : Any = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Union[str, Any]: a : Any = self.feat_extract_dict a : Optional[Any] = True a : Union[str, Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : Any = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : int = feat_extract.model_input_names[0] a : List[Any] = BatchFeature({input_name: speech_inputs} ) a : Union[str, Any] = feat_extract.num_mel_bins # hack! a : Dict = feat_extract.pad(lowerCAmelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCAmelCase__ ) def __a ( self ) -> Union[str, Any]: a : Tuple = self.feat_extract_dict a : str = True a : Optional[Any] = self.feature_extraction_class(**lowerCAmelCase__ ) a : List[Any] = self.feat_extract_tester.prepare_inputs_for_target() a : Dict = [len(lowerCAmelCase__ ) for x in speech_inputs] a : Optional[Any] = feat_extract.model_input_names[0] a : str = BatchFeature({input_name: speech_inputs} ) a : Optional[Any] = min(lowerCAmelCase__ ) a : List[Any] = feat_extract.num_mel_bins # hack! a : Any = feat_extract.pad( lowerCAmelCase__ , padding="max_length" , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCAmelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: from datasets import load_dataset a : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech a : Optional[Any] = ds.sort("id" ).select(range(lowerCAmelCase__ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __a ( self ) -> Union[str, Any]: # fmt: off a : List[Any] = torch.tensor( [2.3_804E-03, 2.0_752E-03, 1.9_836E-03, 2.1_057E-03, 1.6_174E-03, 3.0_518E-04, 9.1_553E-05, 3.3_569E-04, 9.7_656E-04, 1.8_311E-03, 2.0_142E-03, 2.1_057E-03, 1.7_395E-03, 4.5_776E-04, -3.9_673E-04, 4.5_776E-04, 1.0_071E-03, 9.1_553E-05, 4.8_828E-04, 1.1_597E-03, 7.3_242E-04, 9.4_604E-04, 1.8_005E-03, 1.8_311E-03, 8.8_501E-04, 4.2_725E-04, 4.8_828E-04, 7.3_242E-04, 1.0_986E-03, 2.1_057E-03] ) # fmt: on a : List[str] = self._load_datasamples(1 ) a : Union[str, Any] = SpeechTaFeatureExtractor() a : str = feature_extractor(lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 9_3680) ) self.assertTrue(torch.allclose(input_values[0, :30] , lowerCAmelCase__ , atol=1E-6 ) ) def __a ( self ) -> Union[str, Any]: # fmt: off a : Tuple = torch.tensor( [-2.6_870, -3.0_104, -3.1_356, -3.5_352, -3.0_044, -3.0_353, -3.4_719, -3.6_777, -3.1_520, -2.9_435, -2.6_553, -2.8_795, -2.9_944, -2.5_921, -3.0_279, -3.0_386, -3.0_864, -3.1_291, -3.2_353, -2.7_444, -2.6_831, -2.7_287, -3.1_761, -3.1_571, -3.2_726, -3.0_582, -3.1_007, -3.4_533, -3.4_695, -3.0_998] ) # fmt: on a : Dict = self._load_datasamples(1 ) a : Tuple = SpeechTaFeatureExtractor() a : Optional[int] = feature_extractor(audio_target=lowerCAmelCase__ , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , lowerCAmelCase__ , atol=1E-4 ) )
31
1
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class __UpperCamelCase : def __init__( self , lowerCAmelCase__ ) -> None: a : Dict = value a : Node | None = None a : Node | None = None class __UpperCamelCase : def __init__( self , lowerCAmelCase__ ) -> None: a : int = tree def __a ( self , lowerCAmelCase__ ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] ) ->int: '''simple docstring''' a : int = {} a : Union[str, Any] = tokenizer(example["content"] , truncation=_lowercase )["input_ids"] a : Any = len(example["content"] ) / len(output["input_ids"] ) return output a : int = HfArgumentParser(PretokenizationArguments) a : Optional[int] = parser.parse_args() if args.num_workers is None: a : Tuple = multiprocessing.cpu_count() a : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) a : Dict = time.time() a : Tuple = load_dataset(args.dataset_name, split='''train''') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a : Dict = time.time() a : Tuple = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a : Tuple = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
31
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float , _lowercase : float ) ->tuple: '''simple docstring''' a : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a : List[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a : Union[str, Any] = TaTokenizerFast a : Union[str, Any] = {'''configuration_mt5''': ['''MT5Config''', '''MT5OnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ '''MT5EncoderModel''', '''MT5ForConditionalGeneration''', '''MT5ForQuestionAnswering''', '''MT5Model''', '''MT5PreTrainedModel''', '''MT5Stack''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''TFMT5EncoderModel''', '''TFMT5ForConditionalGeneration''', '''TFMT5Model'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['''FlaxMT5EncoderModel''', '''FlaxMT5ForConditionalGeneration''', '''FlaxMT5Model'''] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a : List[Any] = _LazyModule( __name__, globals()['''__file__'''], _import_structure, extra_objects={'''MT5Tokenizer''': MTaTokenizer, '''MT5TokenizerFast''': MTaTokenizerFast}, module_spec=__spec__, )
31
1
"""simple docstring""" import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def _SCREAMING_SNAKE_CASE ( _lowercase : List[Any] ) ->List[str]: # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' with parallel_backend("spark" ): assert ParallelBackendConfig.backend_name == "spark" a : str = [1, 2, 3] with pytest.raises(_lowercase ): with parallel_backend("unsupported backend" ): map_nested(_lowercase , _lowercase , num_proc=2 ) with pytest.raises(_lowercase ): with parallel_backend("unsupported backend" ): map_nested(_lowercase , _lowercase , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize("num_proc" , [2, -1] ) def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->str: '''simple docstring''' a : Dict = [1, 2] a : Tuple = {"a": 1, "b": 2} a : List[Any] = {"a": [1, 2], "b": [3, 4]} a : int = {"a": {"1": 1}, "b": 2} a : Dict = {"a": 1, "b": 2, "c": 3, "d": 4} a : Dict = [2, 3] a : List[str] = {"a": 2, "b": 3} a : Any = {"a": [2, 3], "b": [4, 5]} a : Optional[int] = {"a": {"1": 2}, "b": 3} a : int = {"a": 2, "b": 3, "c": 4, "d": 5} with parallel_backend("spark" ): assert map_nested(_lowercase , _lowercase , num_proc=_lowercase ) == expected_map_nested_sa assert map_nested(_lowercase , _lowercase , num_proc=_lowercase ) == expected_map_nested_sa assert map_nested(_lowercase , _lowercase , num_proc=_lowercase ) == expected_map_nested_sa assert map_nested(_lowercase , _lowercase , num_proc=_lowercase ) == expected_map_nested_sa assert map_nested(_lowercase , _lowercase , num_proc=_lowercase ) == expected_map_nested_sa
31
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 10 , _lowercase : int = 1000 , _lowercase : bool = True ) ->int: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->int: '''simple docstring''' return int((number_a + number_a) / 2 ) def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int , _lowercase : int ) ->None: '''simple docstring''' assert ( isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) and isinstance(_lowercase , _lowercase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(_lowercase : int ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) a : Optional[Any] = lower a : List[Any] = higher a : Tuple = [] while True: a : List[Any] = get_avg(_lowercase , _lowercase ) last_numbers.append(_lowercase ) if answer(_lowercase ) == "low": a : Optional[int] = number elif answer(_lowercase ) == "high": a : Tuple = number else: break print(F"""guess the number : {last_numbers[-1]}""" ) print(F"""details : {last_numbers!s}""" ) def _SCREAMING_SNAKE_CASE ( ) ->None: '''simple docstring''' a : Tuple = int(input("Enter lower value : " ).strip() ) a : Dict = int(input("Enter high value : " ).strip() ) a : Optional[int] = int(input("Enter value to guess : " ).strip() ) guess_the_number(_lowercase , _lowercase , _lowercase ) if __name__ == "__main__": main()
31
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : Optional[int] ) ->List[Any]: '''simple docstring''' a : Union[str, Any] = [1] for i in range(2 , _lowercase ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" a : Dict = [] a : Optional[int] = list(range(_lowercase ) ) # Find permutation while factorials: a : Optional[int] = factorials.pop() a, a : Dict = divmod(_lowercase , _lowercase ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
31
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort a : Any = logging.get_logger(__name__) a : Tuple = { '''tensor(bool)''': np.bool_, '''tensor(int8)''': np.inta, '''tensor(uint8)''': np.uinta, '''tensor(int16)''': np.intaa, '''tensor(uint16)''': np.uintaa, '''tensor(int32)''': np.intaa, '''tensor(uint32)''': np.uintaa, '''tensor(int64)''': np.intaa, '''tensor(uint64)''': np.uintaa, '''tensor(float16)''': np.floataa, '''tensor(float)''': np.floataa, '''tensor(double)''': np.floataa, } class __UpperCamelCase : def __init__( self , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> str: logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) a : Optional[int] = model a : int = kwargs.get("model_save_dir" , lowerCAmelCase__ ) a : Tuple = kwargs.get("latest_model_name" , lowerCAmelCase__ ) def __call__( self , **lowerCAmelCase__ ) -> Dict: a : List[str] = {k: np.array(lowerCAmelCase__ ) for k, v in kwargs.items()} return self.model.run(lowerCAmelCase__ , lowerCAmelCase__ ) @staticmethod def __a ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> Union[str, Any]: if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) a : List[str] = "CPUExecutionProvider" return ort.InferenceSession(lowerCAmelCase__ , providers=[provider] , sess_options=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__ ) -> int: a : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME a : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) a : List[str] = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) a : str = self.model_save_dir.joinpath(lowerCAmelCase__ ) if src_path.exists(): a : Any = Path(lowerCAmelCase__ ).joinpath(lowerCAmelCase__ ) try: shutil.copyfile(lowerCAmelCase__ , lowerCAmelCase__ ) except shutil.SameFileError: pass def __a ( self , lowerCAmelCase__ , **lowerCAmelCase__ , ) -> str: if os.path.isfile(lowerCAmelCase__ ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # saving model weights/files self._save_pretrained(lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> Optional[int]: a : Tuple = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(lowerCAmelCase__ ): a : Tuple = OnnxRuntimeModel.load_model( os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) a : Tuple = Path(lowerCAmelCase__ ) # load model from hub else: # download model a : Optional[Any] = hf_hub_download( repo_id=lowerCAmelCase__ , filename=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , ) a : Optional[int] = Path(lowerCAmelCase__ ).parent a : List[Any] = Path(lowerCAmelCase__ ).name a : int = OnnxRuntimeModel.load_model(lowerCAmelCase__ , provider=lowerCAmelCase__ , sess_options=lowerCAmelCase__ ) return cls(model=lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __a ( cls , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> List[str]: a : Any = None if len(str(lowerCAmelCase__ ).split("@" ) ) == 2: a, a : Tuple = model_id.split("@" ) return cls._from_pretrained( model_id=lowerCAmelCase__ , revision=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , **lowerCAmelCase__ , )
31
1
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position a : List[str] = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip a : Optional[Any] = concatenate_datasets a : Optional[int] = DownloadConfig a : Union[str, Any] = DownloadManager a : Dict = DownloadMode a : int = DownloadConfig a : Optional[int] = DownloadMode a : Tuple = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
31
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : Optional[Any] ) ->str: '''simple docstring''' a : Union[str, Any] = 1.5 a : List[str] = int(factor * num_class_images ) a : Optional[Any] = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=_lowercase ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: a : List[Any] = client.query(text=_lowercase ) if len(_lowercase ) >= factor * num_class_images or num_images > 1E4: break else: a : Optional[int] = int(factor * num_images ) a : str = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=_lowercase , aesthetic_weight=0.1 , ) a : Optional[int] = 0 a : str = 0 a : Any = tqdm(desc="downloading real regularization images" , total=_lowercase ) with open(F"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(F"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( F"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: a : Optional[Any] = class_images[count] count += 1 try: a : str = requests.get(images["url"] ) if img.status_code == 200: a : int = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Optional[int] = argparse.ArgumentParser("" , add_help=_lowercase ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=_lowercase , type=_lowercase ) parser.add_argument("--class_data_dir" , help="path to save images" , required=_lowercase , type=_lowercase ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=_lowercase ) return parser.parse_args() if __name__ == "__main__": a : List[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
31
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __UpperCamelCase ( metaclass=a__ ): lowerCamelCase : Optional[Any] =["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Union[str, Any]: requires_backends(self , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: requires_backends(cls , ["transformers", "torch", "note_seq"] ) @classmethod def __a ( cls , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> int: requires_backends(cls , ["transformers", "torch", "note_seq"] )
31
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file a : Optional[int] = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _SCREAMING_SNAKE_CASE ( _lowercase : Any=None ) ->Optional[Any]: '''simple docstring''' if subparsers is not None: a : int = subparsers.add_parser("tpu-config" , description=_description ) else: a : List[Any] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments a : Dict = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_lowercase , default=_lowercase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_lowercase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_lowercase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) a : Any = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_lowercase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_lowercase ) return parser def _SCREAMING_SNAKE_CASE ( _lowercase : Any ) ->Tuple: '''simple docstring''' a : Union[str, Any] = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_lowercase ): a : Optional[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: a : int = defaults.command_file if not args.command and defaults.commands is not None: a : Union[str, Any] = defaults.commands if not args.tpu_name: a : int = defaults.tpu_name if not args.tpu_zone: a : Union[str, Any] = defaults.tpu_zone if args.accelerate_version == "dev": a : int = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": a : Optional[Any] = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _lowercase ): a : Optional[Any] = F"""accelerate=={args.accelerate_version}""" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: a : int = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _lowercase ): a : Union[str, Any] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate a : Tuple = ["cd /usr/share"] if args.install_accelerate: new_cmd += [F"""pip install {args.accelerate_version}"""] new_cmd += args.command a : List[Any] = "; ".join(_lowercase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess a : str = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"""Running {' '.join(_lowercase )}""" ) return subprocess.run(_lowercase ) print("Successfully setup pod." ) def _SCREAMING_SNAKE_CASE ( ) ->Tuple: '''simple docstring''' a : List[Any] = tpu_command_parser() a : Optional[int] = parser.parse_args() tpu_command_launcher(_lowercase )
31
1
"""simple docstring""" a : List[str] = tuple[float, float, float] a : Union[str, Any] = tuple[float, float, float] def _SCREAMING_SNAKE_CASE ( _lowercase : Pointad , _lowercase : Pointad ) ->Vectorad: '''simple docstring''' a : Tuple = end_pointa[0] - end_pointa[0] a : Dict = end_pointa[1] - end_pointa[1] a : List[str] = end_pointa[2] - end_pointa[2] return (x, y, z) def _SCREAMING_SNAKE_CASE ( _lowercase : Vectorad , _lowercase : Vectorad ) ->Vectorad: '''simple docstring''' a : List[Any] = ab[1] * ac[2] - ab[2] * ac[1] # *i a : Any = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j a : int = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def _SCREAMING_SNAKE_CASE ( _lowercase : Vectorad , _lowercase : int ) ->bool: '''simple docstring''' return tuple(round(_lowercase , _lowercase ) for x in vector ) == (0, 0, 0) def _SCREAMING_SNAKE_CASE ( _lowercase : Pointad , _lowercase : Pointad , _lowercase : Pointad , _lowercase : int = 10 ) ->bool: '''simple docstring''' a : str = create_vector(_lowercase , _lowercase ) a : List[Any] = create_vector(_lowercase , _lowercase ) return is_zero_vector(get_ad_vectors_cross(_lowercase , _lowercase ) , _lowercase )
31
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: a : Tuple = None a : int = logging.get_logger(__name__) a : int = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} a : Optional[int] = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } a : int = { '''facebook/nllb-large-en-ro''': 1024, '''facebook/nllb-200-distilled-600M''': 1024, } # fmt: off a : List[Any] = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[Any] =VOCAB_FILES_NAMES lowerCamelCase : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Dict =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] =["""input_ids""", """attention_mask"""] lowerCamelCase : Union[str, Any] =NllbTokenizer lowerCamelCase : List[int] =[] lowerCamelCase : List[int] =[] def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[Any]: # Mask token behave like a normal word, i.e. include the space before it a : Dict = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token a : Optional[Any] = legacy_behaviour super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) a : int = vocab_file a : Any = False if not self.vocab_file else True a : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) a : str = { lang_code: self.convert_tokens_to_ids(lowerCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } a : List[Any] = src_lang if src_lang is not None else "eng_Latn" a : str = self.convert_tokens_to_ids(self._src_lang ) a : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __a ( self ) -> str: return self._src_lang @src_lang.setter def __a ( self , lowerCAmelCase__ ) -> None: a : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: a : str = [self.sep_token_id] a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) a : Dict = src_lang a : int = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Dict = self.convert_tokens_to_ids(lowerCAmelCase__ ) a : Any = tgt_lang_id return inputs def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: a : Optional[int] = src_lang a : int = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __a ( self ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def __a ( self ) -> str: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __a ( self , lowerCAmelCase__ ) -> None: a : int = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Tuple = [] a : List[str] = [self.eos_token_id, self.cur_lang_code] else: a : int = [self.cur_lang_code] a : int = [self.eos_token_id] a : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) a : Any = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ ) -> None: a : str = self.convert_tokens_to_ids(lowerCAmelCase__ ) if self.legacy_behaviour: a : Optional[Any] = [] a : int = [self.eos_token_id, self.cur_lang_code] else: a : List[Any] = [self.cur_lang_code] a : List[Any] = [self.eos_token_id] a : int = self.convert_ids_to_tokens(self.prefix_tokens ) a : int = self.convert_ids_to_tokens(self.suffix_tokens ) a : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return a : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
31
1
"""simple docstring""" import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel a : Optional[Any] = False a : List[Any] = True a : List[Any] = False if __name__ == "__main__": a : Tuple = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') a : Tuple = parser.parse_args() a : List[str] = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } a : Optional[int] = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } a : Union[str, Any] = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: a : int = reader.read() a : List[Any] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): a : List[Any] = UNetaDModel(**config) else: a : Dict = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel a : List[str] = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) a : str = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: a : str = config[key] del config[key] a : int = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] a : Tuple = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: a : List[Any] = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) a : int = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue a : List[Any] = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: a : Union[str, Any] = param_value a : str = True if not has_changed: a : List[Any] = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
31
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __UpperCamelCase ( a__ ): lowerCamelCase : torch.FloatTensor lowerCamelCase : torch.FloatTensor lowerCamelCase : Optional[torch.FloatTensor] =None class __UpperCamelCase ( a__ , a__ ): lowerCamelCase : Tuple =2 @register_to_config def __init__( self , lowerCAmelCase__ = 0.02 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 1.007 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 0.05 , lowerCAmelCase__ = 50 , ) -> Union[str, Any]: # standard deviation of the initial noise distribution a : Tuple = sigma_max # setable values a : int = None a : np.IntTensor = None a : torch.FloatTensor = None # sigma(t_i) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> torch.FloatTensor: return sample def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[str]: a : List[Any] = num_inference_steps a : List[str] = np.arange(0 , self.num_inference_steps )[::-1].copy() a : int = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) a : List[str] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] a : Any = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa , device=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: a : str = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: a : Dict = 0 # sample eps ~ N(0, S_noise^2 * I) a : Union[str, Any] = self.config.s_noise * randn_tensor(sample.shape , generator=lowerCAmelCase__ ).to(sample.device ) a : Any = sigma + gamma * sigma a : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Union[str, Any] = sample_hat + sigma_hat * model_output a : Tuple = (sample_hat - pred_original_sample) / sigma_hat a : List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[KarrasVeOutput, Tuple]: a : Optional[int] = sample_prev + sigma_prev * model_output a : str = (sample_prev - pred_original_sample) / sigma_prev a : Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowerCAmelCase__ , derivative=lowerCAmelCase__ , pred_original_sample=lowerCAmelCase__ ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: raise NotImplementedError()
31
1
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __UpperCamelCase ( unittest.TestCase ): def __a ( self ) -> Optional[Any]: a : Optional[int] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : str = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : List[str] = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> List[Any]: a : Optional[Any] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Tuple: a : Tuple = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ ) ) def __a ( self ) -> Dict: a : Dict = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Dict = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: a : List[Any] = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Any = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> int: # pass variant but use the non-variant filenames a : int = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] a : Tuple = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : str = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a : Any = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> str: a : Union[str, Any] = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> List[str]: # pass variant but use the non-variant filenames a : Optional[int] = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] a : str = "fp16" self.assertTrue(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) ) def __a ( self ) -> Optional[Any]: a : Any = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] a : Optional[int] = "fp16" self.assertFalse(is_safetensors_compatible(lowerCAmelCase__ , variant=lowerCAmelCase__ ) )
31
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a : Optional[Any] = datasets.utils.logging.get_logger(__name__) a : Union[str, Any] = ['''names''', '''prefix'''] a : Any = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] a : Any = ['''encoding_errors''', '''on_bad_lines'''] a : List[str] = ['''date_format'''] @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): lowerCamelCase : str ="," lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[Union[int, List[int], str]] ="infer" lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[List[str]] =None lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] =None lowerCamelCase : Optional[Union[List[int], List[str]]] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] =None lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] =None lowerCamelCase : Optional[list] =None lowerCamelCase : Optional[list] =None lowerCamelCase : bool =False lowerCamelCase : Optional[Union[int, List[int]]] =None lowerCamelCase : Optional[int] =None lowerCamelCase : Optional[Union[str, List[str]]] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : bool =True lowerCamelCase : Optional[str] =None lowerCamelCase : str ="." lowerCamelCase : Optional[str] =None lowerCamelCase : str ='"' lowerCamelCase : int =0 lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : Optional[str] =None lowerCamelCase : bool =True lowerCamelCase : bool =True lowerCamelCase : int =0 lowerCamelCase : bool =True lowerCamelCase : bool =False lowerCamelCase : Optional[str] =None lowerCamelCase : int =1_0000 lowerCamelCase : Optional[datasets.Features] =None lowerCamelCase : Optional[str] ="strict" lowerCamelCase : Literal["error", "warn", "skip"] ="error" lowerCamelCase : Optional[str] =None def __a ( self ) -> Dict: if self.delimiter is not None: a : int = self.delimiter if self.column_names is not None: a : Any = self.column_names @property def __a ( self ) -> List[str]: a : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __UpperCamelCase ( datasets.ArrowBasedBuilder ): lowerCamelCase : Union[str, Any] =CsvConfig def __a ( self ) -> Optional[Any]: return datasets.DatasetInfo(features=self.config.features ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) a : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): a : Tuple = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Tuple = [files] a : int = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] a : int = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a : Any = [files] a : List[str] = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def __a ( self , lowerCAmelCase__ ) -> pa.Table: if self.config.features is not None: a : Optional[Any] = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast a : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example a : Union[str, Any] = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def __a ( self , lowerCAmelCase__ ) -> Any: a : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str a : Any = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): a : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): a : Any = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(lowerCAmelCase__ )}: {e}""" ) raise
31
1