code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging lowerCAmelCase_ : Any = logging.get_logger(__name__) class UpperCamelCase_ ( a_ ): _A : List[Any] = ['input_features', 'is_longer'] def __init__( self , snake_case__=64 , snake_case__=4_80_00 , snake_case__=4_80 , snake_case__=10 , snake_case__=10_24 , snake_case__=0.0 , snake_case__=False , snake_case__ = 0 , snake_case__ = 1_40_00 , snake_case__ = None , snake_case__ = "fusion" , snake_case__ = "repeatpad" , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__( feature_size=snake_case__ , sampling_rate=snake_case__ , padding_value=snake_case__ , return_attention_mask=snake_case__ , **snake_case__ , ) UpperCAmelCase = top_db UpperCAmelCase = truncation UpperCAmelCase = padding UpperCAmelCase = fft_window_size UpperCAmelCase = (fft_window_size >> 1) + 1 UpperCAmelCase = hop_length UpperCAmelCase = max_length_s UpperCAmelCase = max_length_s * sampling_rate UpperCAmelCase = sampling_rate UpperCAmelCase = frequency_min UpperCAmelCase = frequency_max UpperCAmelCase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=snake_case__ , min_frequency=snake_case__ , max_frequency=snake_case__ , sampling_rate=snake_case__ , norm=snake_case__ , mel_scale="""htk""" , ) UpperCAmelCase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=snake_case__ , min_frequency=snake_case__ , max_frequency=snake_case__ , sampling_rate=snake_case__ , norm="""slaney""" , mel_scale="""slaney""" , ) def UpperCamelCase_ ( self ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase = copy.deepcopy(self.__dict__ ) UpperCAmelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> np.ndarray: """simple docstring""" UpperCAmelCase = spectrogram( snake_case__ , window_function(self.fft_window_size , """hann""" ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=snake_case__ , log_mel="""dB""" , ) return log_mel_spectrogram.T def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Tuple: """simple docstring""" UpperCAmelCase = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCAmelCase = [0] # randomly choose index for each part UpperCAmelCase = np.random.choice(ranges[0] ) UpperCAmelCase = np.random.choice(ranges[1] ) UpperCAmelCase = np.random.choice(ranges[2] ) UpperCAmelCase = mel[idx_front : idx_front + chunk_frames, :] UpperCAmelCase = mel[idx_middle : idx_middle + chunk_frames, :] UpperCAmelCase = mel[idx_back : idx_back + chunk_frames, :] UpperCAmelCase = torch.tensor(mel[None, None, :] ) UpperCAmelCase = torch.nn.functional.interpolate( snake_case__ , size=[chunk_frames, 64] , mode="""bilinear""" , align_corners=snake_case__ ) UpperCAmelCase = mel_shrink[0][0].numpy() UpperCAmelCase = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> np.array: """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCAmelCase = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCAmelCase = len(snake_case__ ) - max_length UpperCAmelCase = np.random.randint(0 , overflow + 1 ) UpperCAmelCase = waveform[idx : idx + max_length] UpperCAmelCase = self._np_extract_fbank_features(snake_case__ , self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCAmelCase = self._np_extract_fbank_features(snake_case__ , self.mel_filters ) UpperCAmelCase = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCAmelCase = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCAmelCase = np.stack([mel, mel, mel, mel] , axis=0 ) UpperCAmelCase = False else: UpperCAmelCase = self._random_mel_fusion(snake_case__ , snake_case__ , snake_case__ ) UpperCAmelCase = True else: raise NotImplementedError(f'''data_truncating {truncation} not implemented''' ) else: UpperCAmelCase = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCAmelCase = int(max_length / len(snake_case__ ) ) UpperCAmelCase = np.stack(np.tile(snake_case__ , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCAmelCase = int(max_length / len(snake_case__ ) ) UpperCAmelCase = np.stack(np.tile(snake_case__ , snake_case__ ) ) UpperCAmelCase = np.pad(snake_case__ , (0, max_length - waveform.shape[0]) , mode="""constant""" , constant_values=0 ) if truncation == "fusion": UpperCAmelCase = self._np_extract_fbank_features(snake_case__ , self.mel_filters ) UpperCAmelCase = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: UpperCAmelCase = self._np_extract_fbank_features(snake_case__ , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , **snake_case__ , ) -> BatchFeature: """simple docstring""" UpperCAmelCase = truncation if truncation is not None else self.truncation UpperCAmelCase = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' f''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' f''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCAmelCase = isinstance(snake_case__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) UpperCAmelCase = is_batched_numpy or ( isinstance(snake_case__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase = [np.asarray(snake_case__ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(snake_case__ , np.ndarray ): UpperCAmelCase = np.asarray(snake_case__ , dtype=np.floataa ) elif isinstance(snake_case__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCAmelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase = [np.asarray(snake_case__ )] # convert to mel spectrogram, truncate and pad if needed. UpperCAmelCase = [ self._get_input_mel(snake_case__ , max_length if max_length else self.nb_max_samples , snake_case__ , snake_case__ ) for waveform in raw_speech ] UpperCAmelCase = [] UpperCAmelCase = [] for mel, longer in padded_inputs: input_mel.append(snake_case__ ) is_longer.append(snake_case__ ) if truncation == "fusion" and sum(snake_case__ ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCAmelCase = np.random.randint(0 , len(snake_case__ ) ) UpperCAmelCase = True if isinstance(input_mel[0] , snake_case__ ): UpperCAmelCase = [np.asarray(snake_case__ , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCAmelCase = [[longer] for longer in is_longer] UpperCAmelCase = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCAmelCase = BatchFeature(snake_case__ ) if return_tensors is not None: UpperCAmelCase = input_features.convert_to_tensors(snake_case__ ) return input_features
673
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = 32 , snake_case__ = True , snake_case__ = 1 / 2_55 , snake_case__ = True , snake_case__ = True , snake_case__ = [0.48_145_466, 0.4_578_275, 0.40_821_073] , snake_case__ = [0.26_862_954, 0.26_130_258, 0.27_577_711] , snake_case__ = True , snake_case__=7 , snake_case__=30 , snake_case__=4_00 , snake_case__=3 , ) -> List[str]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = do_resize UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_88} UpperCAmelCase = size_divisor UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_pad UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase_ ( self , snake_case__ , snake_case__=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = image_inputs[0] if isinstance(snake_case__ , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] UpperCAmelCase = size / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size UpperCAmelCase = int((13_33 / 8_00) * size ) if max(snake_case__ , snake_case__ ) > max_size: UpperCAmelCase = max_size / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase , UpperCAmelCase = int(newh + 0.5 ), int(neww + 0.5 ) UpperCAmelCase , UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """size_divisor""" ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
673
1
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule lowerCAmelCase_ : Optional[Any] = { '''config''': [ '''EXTERNAL_DATA_FORMAT_SIZE_LIMIT''', '''OnnxConfig''', '''OnnxConfigWithPast''', '''OnnxSeq2SeqConfigWithPast''', '''PatchingSpec''', ], '''convert''': ['''export''', '''validate_model_outputs'''], '''features''': ['''FeaturesManager'''], '''utils''': ['''ParameterFormat''', '''compute_serialized_parameters_size'''], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys lowerCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[str] = XLMRobertaTokenizer _A : List[str] = XLMRobertaTokenizerFast _A : Optional[Any] = True _A : List[str] = True def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(snake_case__ ) , 10_02 ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(snake_case__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) UpperCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=snake_case__ ) UpperCAmelCase = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = """I was born in 92000, and this is falsé.""" UpperCAmelCase = tokenizer.tokenize(snake_case__ ) UpperCAmelCase = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
673
1
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase_ : str = (7_2_0, 1_2_8_0) # Height, Width lowerCAmelCase_ : Optional[int] = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase_ : Any = 1 / 1_0_0 lowerCAmelCase_ : List[str] = '''''' lowerCAmelCase_ : Union[str, Any] = '''''' lowerCAmelCase_ : Dict = '''''' lowerCAmelCase_ : Optional[int] = 2_5_0 def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = get_dataset(lowerCAmelCase , lowerCAmelCase ) for index in range(lowerCAmelCase ): UpperCAmelCase = random.sample(range(len(lowerCAmelCase ) ) , 4 ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = update_image_and_anno( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , filter_scale=lowerCAmelCase , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' UpperCAmelCase = random_chars(32 ) UpperCAmelCase = path.split(os.sep )[-1].rsplit(""".""" , 1 )[0] UpperCAmelCase = F'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(F'''{file_root}.jpg''' , lowerCAmelCase , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) UpperCAmelCase = [] for anno in new_annos: UpperCAmelCase = anno[3] - anno[1] UpperCAmelCase = anno[4] - anno[2] UpperCAmelCase = anno[1] + width / 2 UpperCAmelCase = anno[2] + height / 2 UpperCAmelCase = F'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(lowerCAmelCase ) with open(F'''{file_root}.txt''' , """w""" ) as outfile: outfile.write("""\n""".join(line for line in annos_list ) ) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [] UpperCAmelCase = [] for label_file in glob.glob(os.path.join(lowerCAmelCase , """*.txt""" ) ): UpperCAmelCase = label_file.split(os.sep )[-1].rsplit(""".""" , 1 )[0] with open(lowerCAmelCase ) as in_file: UpperCAmelCase = in_file.readlines() UpperCAmelCase = os.path.join(lowerCAmelCase , F'''{label_name}.jpg''' ) UpperCAmelCase = [] for obj_list in obj_lists: UpperCAmelCase = obj_list.rstrip("""\n""" ).split(""" """ ) UpperCAmelCase = float(obj[1] ) - float(obj[3] ) / 2 UpperCAmelCase = float(obj[2] ) - float(obj[4] ) / 2 UpperCAmelCase = float(obj[1] ) + float(obj[3] ) / 2 UpperCAmelCase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(lowerCAmelCase ) labels.append(lowerCAmelCase ) return img_paths, labels def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 0.0 , ): '''simple docstring''' UpperCAmelCase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) UpperCAmelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCAmelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCAmelCase = int(scale_x * output_size[1] ) UpperCAmelCase = int(scale_y * output_size[0] ) UpperCAmelCase = [] UpperCAmelCase = [] for i, index in enumerate(lowerCAmelCase ): UpperCAmelCase = all_img_list[index] path_list.append(lowerCAmelCase ) UpperCAmelCase = all_annos[index] UpperCAmelCase = cva.imread(lowerCAmelCase ) if i == 0: # top-left UpperCAmelCase = cva.resize(lowerCAmelCase , (divid_point_x, divid_point_y) ) UpperCAmelCase = img for bbox in img_annos: UpperCAmelCase = bbox[1] * scale_x UpperCAmelCase = bbox[2] * scale_y UpperCAmelCase = bbox[3] * scale_x UpperCAmelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right UpperCAmelCase = cva.resize(lowerCAmelCase , (output_size[1] - divid_point_x, divid_point_y) ) UpperCAmelCase = img for bbox in img_annos: UpperCAmelCase = scale_x + bbox[1] * (1 - scale_x) UpperCAmelCase = bbox[2] * scale_y UpperCAmelCase = scale_x + bbox[3] * (1 - scale_x) UpperCAmelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left UpperCAmelCase = cva.resize(lowerCAmelCase , (divid_point_x, output_size[0] - divid_point_y) ) UpperCAmelCase = img for bbox in img_annos: UpperCAmelCase = bbox[1] * scale_x UpperCAmelCase = scale_y + bbox[2] * (1 - scale_y) UpperCAmelCase = bbox[3] * scale_x UpperCAmelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right UpperCAmelCase = cva.resize( lowerCAmelCase , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) UpperCAmelCase = img for bbox in img_annos: UpperCAmelCase = scale_x + bbox[1] * (1 - scale_x) UpperCAmelCase = scale_y + bbox[2] * (1 - scale_y) UpperCAmelCase = scale_x + bbox[3] * (1 - scale_x) UpperCAmelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: UpperCAmelCase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" UpperCAmelCase = ascii_lowercase + digits return "".join(random.choice(lowerCAmelCase ) for _ in range(lowerCAmelCase ) ) if __name__ == "__main__": main() print('''DONE ✅''')
673
"""simple docstring""" import socket def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase = socket.gethostname() UpperCAmelCase = 12312 sock.connect((host, port) ) sock.send(b"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: UpperCAmelCase = sock.recv(1024 ) if not data: break out_file.write(lowerCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
673
1
"""simple docstring""" import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowerCAmelCase_ : Dict = logging.get_logger(__name__) lowerCAmelCase_ : Optional[Any] = { '''microsoft/conditional-detr-resnet-50''': ( '''https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json''' ), } class UpperCamelCase_ ( a_ ): _A : List[Any] = 'conditional_detr' _A : Any = ['past_key_values'] _A : Union[str, Any] = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , snake_case__=True , snake_case__=None , snake_case__=3 , snake_case__=3_00 , snake_case__=6 , snake_case__=20_48 , snake_case__=8 , snake_case__=6 , snake_case__=20_48 , snake_case__=8 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=True , snake_case__="relu" , snake_case__=2_56 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.02 , snake_case__=1.0 , snake_case__=False , snake_case__="sine" , snake_case__="resnet50" , snake_case__=True , snake_case__=False , snake_case__=2 , snake_case__=5 , snake_case__=2 , snake_case__=1 , snake_case__=1 , snake_case__=2 , snake_case__=5 , snake_case__=2 , snake_case__=0.25 , **snake_case__ , ) -> List[Any]: """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = backbone_config.get("""model_type""" ) UpperCAmelCase = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase = config_class.from_dict(snake_case__ ) UpperCAmelCase = use_timm_backbone UpperCAmelCase = backbone_config UpperCAmelCase = num_channels UpperCAmelCase = num_queries UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = init_xavier_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = encoder_layers UpperCAmelCase = auxiliary_loss UpperCAmelCase = position_embedding_type UpperCAmelCase = backbone UpperCAmelCase = use_pretrained_backbone UpperCAmelCase = dilation # Hungarian matcher UpperCAmelCase = class_cost UpperCAmelCase = bbox_cost UpperCAmelCase = giou_cost # Loss coefficients UpperCAmelCase = mask_loss_coefficient UpperCAmelCase = dice_loss_coefficient UpperCAmelCase = cls_loss_coefficient UpperCAmelCase = bbox_loss_coefficient UpperCAmelCase = giou_loss_coefficient UpperCAmelCase = focal_alpha super().__init__(is_encoder_decoder=snake_case__ , **snake_case__ ) @property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return self.d_model def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCAmelCase = self.backbone_config.to_dict() UpperCAmelCase = self.__class__.model_type return output class UpperCamelCase_ ( a_ ): _A : Union[str, Any] = version.parse('1.11' ) @property def UpperCamelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase_ ( self ) -> float: """simple docstring""" return 1e-5 @property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return 12
673
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
1
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib lowerCAmelCase_ : List[str] = get_logger() lowerCAmelCase_ : Optional[dict] = None class UpperCamelCase_ ( TensorFormatter[Mapping, 'jax.Array', Mapping] ): def __init__( self , snake_case__=None , snake_case__=None , **snake_case__ ) -> int: """simple docstring""" super().__init__(features=snake_case__ ) import jax from jaxlib.xla_client import Device if isinstance(snake_case__ , snake_case__ ): raise ValueError( f'''Expected {device} to be a `str` not {type(snake_case__ )}, as `jaxlib.xla_extension.Device` ''' """is not serializable neither with `pickle` nor with `dill`. Instead you can surround """ """the device with `str()` to get its string identifier that will be internally mapped """ """to the actual `jaxlib.xla_extension.Device`.""" ) UpperCAmelCase = device if isinstance(snake_case__ , snake_case__ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: UpperCAmelCase = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( f'''Device with string identifier {self.device} not listed among the available ''' f'''devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default ''' f'''device: {str(jax.devices()[0] )}.''' ) UpperCAmelCase = str(jax.devices()[0] ) UpperCAmelCase = jnp_array_kwargs @staticmethod def UpperCamelCase_ ( ) -> Dict[str, "jaxlib.xla_extension.Device"]: """simple docstring""" import jax return {str(snake_case__ ): device for device in jax.devices()} def UpperCamelCase_ ( self , snake_case__ ) -> Tuple: """simple docstring""" import jax import jax.numpy as jnp if isinstance(snake_case__ , snake_case__ ) and column: if all( isinstance(snake_case__ , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(snake_case__ , axis=0 ) return column def UpperCamelCase_ ( self , snake_case__ ) -> Dict: """simple docstring""" import jax import jax.numpy as jnp if isinstance(snake_case__ , (str, bytes, type(snake_case__ )) ): return value elif isinstance(snake_case__ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() UpperCAmelCase = {} if isinstance(snake_case__ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: UpperCAmelCase = {"""dtype""": jnp.intaa} else: UpperCAmelCase = {"""dtype""": jnp.intaa} elif isinstance(snake_case__ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): UpperCAmelCase = {"""dtype""": jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(snake_case__ , PIL.Image.Image ): UpperCAmelCase = np.asarray(snake_case__ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: UpperCAmelCase = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(snake_case__ , **{**default_dtype, **self.jnp_array_kwargs} ) def UpperCamelCase_ ( self , snake_case__ ) -> Any: """simple docstring""" import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(snake_case__ , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(snake_case__ , """__array__""" ) and not isinstance(snake_case__ , jax.Array ): UpperCAmelCase = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(snake_case__ , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(snake_case__ ) for substruct in data_struct] ) elif isinstance(snake_case__ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(snake_case__ ) for substruct in data_struct] ) return self._tensorize(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> int: """simple docstring""" return map_nested(self._recursive_tensorize , snake_case__ , map_list=snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Mapping: """simple docstring""" UpperCAmelCase = self.numpy_arrow_extractor().extract_row(snake_case__ ) UpperCAmelCase = self.python_features_decoder.decode_row(snake_case__ ) return self.recursive_tensorize(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> "jax.Array": """simple docstring""" UpperCAmelCase = self.numpy_arrow_extractor().extract_column(snake_case__ ) UpperCAmelCase = self.python_features_decoder.decode_column(snake_case__ , pa_table.column_names[0] ) UpperCAmelCase = self.recursive_tensorize(snake_case__ ) UpperCAmelCase = self._consolidate(snake_case__ ) return column def UpperCamelCase_ ( self , snake_case__ ) -> Mapping: """simple docstring""" UpperCAmelCase = self.numpy_arrow_extractor().extract_batch(snake_case__ ) UpperCAmelCase = self.python_features_decoder.decode_batch(snake_case__ ) UpperCAmelCase = self.recursive_tensorize(snake_case__ ) for column_name in batch: UpperCAmelCase = self._consolidate(batch[column_name] ) return batch
673
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = list(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): UpperCAmelCase = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _lowerCAmelCase ( lowerCAmelCase = None , lowerCAmelCase = 128 ): '''simple docstring''' if function is None: return functools.partial(lowerCAmelCase , starting_batch_size=lowerCAmelCase ) UpperCAmelCase = starting_batch_size def decorator(*lowerCAmelCase , **lowerCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() UpperCAmelCase = list(inspect.signature(lowerCAmelCase ).parameters.keys() ) # Guard against user error if len(lowerCAmelCase ) < (len(lowerCAmelCase ) + 1): UpperCAmelCase = """, """.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) except Exception as e: if should_reduce_batch_size(lowerCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
673
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ : Dict = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Union[str, Any] = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys lowerCAmelCase_ : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
1
"""simple docstring""" import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCAmelCase_ : Tuple = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCAmelCase_ : Any = '''cuda''' if torch.cuda.is_available() else '''cpu''' def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=100 , lowerCAmelCase=" " ): '''simple docstring''' UpperCAmelCase = text.split(lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(lowerCAmelCase ) , lowerCAmelCase )] def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = [], [] for title, text in zip(documents["""title"""] , documents["""text"""] ): if text is not None: for passage in split_text(lowerCAmelCase ): titles.append(title if title is not None else """""" ) texts.append(lowerCAmelCase ) return {"title": titles, "text": texts} def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = ctx_tokenizer( documents["""title"""] , documents["""text"""] , truncation=lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" )["""input_ids"""] UpperCAmelCase = ctx_encoder(input_ids.to(device=lowerCAmelCase ) , return_dict=lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): '''simple docstring''' ###################################### logger.info("""Step 1 - Create the dataset""" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way UpperCAmelCase = load_dataset( """csv""" , data_files=[rag_example_args.csv_path] , split="""train""" , delimiter="""\t""" , column_names=["""title""", """text"""] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words UpperCAmelCase = dataset.map(lowerCAmelCase , batched=lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings UpperCAmelCase = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=lowerCAmelCase ) UpperCAmelCase = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) UpperCAmelCase = Features( {"""text""": Value("""string""" ), """title""": Value("""string""" ), """embeddings""": Sequence(Value("""float32""" ) )} ) # optional, save as float32 instead of float64 to save space UpperCAmelCase = dataset.map( partial(lowerCAmelCase , ctx_encoder=lowerCAmelCase , ctx_tokenizer=lowerCAmelCase ) , batched=lowerCAmelCase , batch_size=processing_args.batch_size , features=lowerCAmelCase , ) # And finally save your dataset UpperCAmelCase = os.path.join(rag_example_args.output_dir , """my_knowledge_dataset""" ) dataset.save_to_disk(lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("""Step 2 - Index the dataset""" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search UpperCAmelCase = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("""embeddings""" , custom_index=lowerCAmelCase ) # And save the index UpperCAmelCase = os.path.join(rag_example_args.output_dir , """my_knowledge_dataset_hnsw_index.faiss""" ) dataset.get_index("""embeddings""" ).save(lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class UpperCamelCase_ : _A : str = field( default=str(Path(a_ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) _A : Optional[str] = field( default=a_ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) _A : str = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) _A : str = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) _A : Optional[str] = field( default=str(Path(a_ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class UpperCamelCase_ : _A : Optional[int] = field( default=a_ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) _A : int = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class UpperCamelCase_ : _A : int = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) _A : int = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCAmelCase_ : List[str] = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : List[Any] = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
673
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer lowerCAmelCase_ : List[str] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase_ : Union[str, Any] = { '''vocab_file''': { '''google/electra-small-generator''': ( '''https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt''' ), '''google/electra-base-generator''': '''https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt''', '''google/electra-large-generator''': ( '''https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt''' ), '''google/electra-small-discriminator''': ( '''https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt''' ), '''google/electra-base-discriminator''': ( '''https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt''' ), '''google/electra-large-discriminator''': ( '''https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''google/electra-small-generator''': ( '''https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json''' ), '''google/electra-base-generator''': ( '''https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json''' ), '''google/electra-large-generator''': ( '''https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json''' ), '''google/electra-small-discriminator''': ( '''https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json''' ), '''google/electra-base-discriminator''': ( '''https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json''' ), '''google/electra-large-discriminator''': ( '''https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase_ : Dict = { '''google/electra-small-generator''': 5_1_2, '''google/electra-base-generator''': 5_1_2, '''google/electra-large-generator''': 5_1_2, '''google/electra-small-discriminator''': 5_1_2, '''google/electra-base-discriminator''': 5_1_2, '''google/electra-large-discriminator''': 5_1_2, } lowerCAmelCase_ : List[str] = { '''google/electra-small-generator''': {'''do_lower_case''': True}, '''google/electra-base-generator''': {'''do_lower_case''': True}, '''google/electra-large-generator''': {'''do_lower_case''': True}, '''google/electra-small-discriminator''': {'''do_lower_case''': True}, '''google/electra-base-discriminator''': {'''do_lower_case''': True}, '''google/electra-large-discriminator''': {'''do_lower_case''': True}, } class UpperCamelCase_ ( a_ ): _A : Tuple = VOCAB_FILES_NAMES _A : int = PRETRAINED_VOCAB_FILES_MAP _A : str = PRETRAINED_INIT_CONFIGURATION _A : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : Optional[Any] = ElectraTokenizer def __init__( self , snake_case__=None , snake_case__=None , snake_case__=True , snake_case__="[UNK]" , snake_case__="[SEP]" , snake_case__="[PAD]" , snake_case__="[CLS]" , snake_case__="[MASK]" , snake_case__=True , snake_case__=None , **snake_case__ , ) -> str: """simple docstring""" super().__init__( snake_case__ , tokenizer_file=snake_case__ , do_lower_case=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , tokenize_chinese_chars=snake_case__ , strip_accents=snake_case__ , **snake_case__ , ) UpperCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , snake_case__ ) != do_lower_case or normalizer_state.get("""strip_accents""" , snake_case__ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , snake_case__ ) != tokenize_chinese_chars ): UpperCAmelCase = getattr(snake_case__ , normalizer_state.pop("""type""" ) ) UpperCAmelCase = do_lower_case UpperCAmelCase = strip_accents UpperCAmelCase = tokenize_chinese_chars UpperCAmelCase = normalizer_class(**snake_case__ ) UpperCAmelCase = do_lower_case def UpperCamelCase_ ( self , snake_case__ , snake_case__=None ) -> int: """simple docstring""" UpperCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase = self._tokenizer.model.save(snake_case__ , name=snake_case__ ) return tuple(snake_case__ )
673
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( a_ ): _A : Optional[int] = 'facebook/bart-large-mnli' _A : Union[str, Any] = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) _A : Dict = 'text_classifier' _A : Union[str, Any] = AutoTokenizer _A : Tuple = AutoModelForSequenceClassification _A : Optional[int] = ['text', ['text']] _A : Dict = ['text'] def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setup() UpperCAmelCase = self.model.config UpperCAmelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = labels return self.pre_processor( [text] * len(snake_case__ ) , [f'''This example is {label}''' for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def UpperCamelCase_ ( self , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = outputs.logits UpperCAmelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
673
1
"""simple docstring""" # limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( '''pipelines_utils''', '''0.22.0''', '''Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.''', standard_warn=False, stacklevel=3, )
673
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase_ ( a_ ): _A : Union[List[PIL.Image.Image], np.ndarray] _A : Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
673
1
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = CTRLTokenizer _A : Any = False _A : Union[str, Any] = False def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase = dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case__ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(snake_case__ ) ) def UpperCamelCase_ ( self , **snake_case__ ) -> Tuple: """simple docstring""" kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Optional[int]: """simple docstring""" UpperCAmelCase = """adapt react readapt apt""" UpperCAmelCase = """adapt react readapt apt""" return input_text, output_text def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCAmelCase = """adapt react readapt apt""" UpperCAmelCase = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase = tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokens + [tokenizer.unk_token] UpperCAmelCase = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case__ ) , snake_case__ )
673
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : Any = { '''configuration_encodec''': [ '''ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EncodecConfig''', ], '''feature_extraction_encodec''': ['''EncodecFeatureExtractor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ '''ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EncodecModel''', '''EncodecPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys lowerCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
1
"""simple docstring""" import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=0.9_99 , lowerCAmelCase="cosine" , ): '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(lowerCAmelCase ): return math.cos((t + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(lowerCAmelCase ): return math.exp(t * -12.0 ) else: raise ValueError(F'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) UpperCAmelCase = [] for i in range(lowerCAmelCase ): UpperCAmelCase = i / num_diffusion_timesteps UpperCAmelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(lowerCAmelCase ) / alpha_bar_fn(lowerCAmelCase ) , lowerCAmelCase ) ) return torch.tensor(lowerCAmelCase , dtype=torch.floataa ) class UpperCamelCase_ ( a_ , a_ ): _A : int = [e.name for e in KarrasDiffusionSchedulers] _A : str = 2 @register_to_config def __init__( self , snake_case__ = 10_00 , snake_case__ = 0.00_085 , snake_case__ = 0.012 , snake_case__ = "linear" , snake_case__ = None , snake_case__ = "epsilon" , snake_case__ = False , snake_case__ = False , snake_case__ = 1.0 , snake_case__ = "linspace" , snake_case__ = 0 , ) -> int: """simple docstring""" if trained_betas is not None: UpperCAmelCase = torch.tensor(snake_case__ , dtype=torch.floataa ) elif beta_schedule == "linear": UpperCAmelCase = torch.linspace(snake_case__ , snake_case__ , snake_case__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. UpperCAmelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , snake_case__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule UpperCAmelCase = betas_for_alpha_bar(snake_case__ , alpha_transform_type="""cosine""" ) elif beta_schedule == "exp": UpperCAmelCase = betas_for_alpha_bar(snake_case__ , alpha_transform_type="""exp""" ) else: raise NotImplementedError(f'''{beta_schedule} does is not implemented for {self.__class__}''' ) UpperCAmelCase = 1.0 - self.betas UpperCAmelCase = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(snake_case__ , snake_case__ , snake_case__ ) UpperCAmelCase = use_karras_sigmas def UpperCamelCase_ ( self , snake_case__ , snake_case__=None ) -> str: """simple docstring""" if schedule_timesteps is None: UpperCAmelCase = self.timesteps UpperCAmelCase = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: UpperCAmelCase = 1 if len(snake_case__ ) > 1 else 0 else: UpperCAmelCase = timestep.cpu().item() if torch.is_tensor(snake_case__ ) else timestep UpperCAmelCase = self._index_counter[timestep_int] return indices[pos].item() @property def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def UpperCamelCase_ ( self , snake_case__ , snake_case__ , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase = self.index_for_timestep(snake_case__ ) UpperCAmelCase = self.sigmas[step_index] UpperCAmelCase = sample / ((sigma**2 + 1) ** 0.5) return sample def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None , snake_case__ = None , ) -> List[Any]: """simple docstring""" UpperCAmelCase = num_inference_steps UpperCAmelCase = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": UpperCAmelCase = np.linspace(0 , num_train_timesteps - 1 , snake_case__ , dtype=snake_case__ )[::-1].copy() elif self.config.timestep_spacing == "leading": UpperCAmelCase = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCAmelCase = (np.arange(0 , snake_case__ ) * step_ratio).round()[::-1].copy().astype(snake_case__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": UpperCAmelCase = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCAmelCase = (np.arange(snake_case__ , 0 , -step_ratio )).round().copy().astype(snake_case__ ) timesteps -= 1 else: raise ValueError( f'''{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.''' ) UpperCAmelCase = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) UpperCAmelCase = np.log(snake_case__ ) UpperCAmelCase = np.interp(snake_case__ , np.arange(0 , len(snake_case__ ) ) , snake_case__ ) if self.config.use_karras_sigmas: UpperCAmelCase = self._convert_to_karras(in_sigmas=snake_case__ , num_inference_steps=self.num_inference_steps ) UpperCAmelCase = np.array([self._sigma_to_t(snake_case__ , snake_case__ ) for sigma in sigmas] ) UpperCAmelCase = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) UpperCAmelCase = torch.from_numpy(snake_case__ ).to(device=snake_case__ ) UpperCAmelCase = torch.cat([sigmas[:1], sigmas[1:-1].repeat_interleave(2 ), sigmas[-1:]] ) UpperCAmelCase = torch.from_numpy(snake_case__ ) UpperCAmelCase = torch.cat([timesteps[:1], timesteps[1:].repeat_interleave(2 )] ) if str(snake_case__ ).startswith("""mps""" ): # mps does not support float64 UpperCAmelCase = timesteps.to(snake_case__ , dtype=torch.floataa ) else: UpperCAmelCase = timesteps.to(device=snake_case__ ) # empty dt and derivative UpperCAmelCase = None UpperCAmelCase = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter UpperCAmelCase = defaultdict(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = np.log(snake_case__ ) # get distribution UpperCAmelCase = log_sigma - log_sigmas[:, np.newaxis] # get sigmas range UpperCAmelCase = np.cumsum((dists >= 0) , axis=0 ).argmax(axis=0 ).clip(max=log_sigmas.shape[0] - 2 ) UpperCAmelCase = low_idx + 1 UpperCAmelCase = log_sigmas[low_idx] UpperCAmelCase = log_sigmas[high_idx] # interpolate sigmas UpperCAmelCase = (low - log_sigma) / (low - high) UpperCAmelCase = np.clip(snake_case__ , 0 , 1 ) # transform interpolation to time range UpperCAmelCase = (1 - w) * low_idx + w * high_idx UpperCAmelCase = t.reshape(sigma.shape ) return t def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase = in_sigmas[-1].item() UpperCAmelCase = in_sigmas[0].item() UpperCAmelCase = 7.0 # 7.0 is the value used in the paper UpperCAmelCase = np.linspace(0 , 1 , snake_case__ ) UpperCAmelCase = sigma_min ** (1 / rho) UpperCAmelCase = sigma_max ** (1 / rho) UpperCAmelCase = (max_inv_rho + ramp * (min_inv_rho - max_inv_rho)) ** rho return sigmas @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.dt is None def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase = self.index_for_timestep(snake_case__ ) # advance index counter by 1 UpperCAmelCase = timestep.cpu().item() if torch.is_tensor(snake_case__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: UpperCAmelCase = self.sigmas[step_index] UpperCAmelCase = self.sigmas[step_index + 1] else: # 2nd order / Heun's method UpperCAmelCase = self.sigmas[step_index - 1] UpperCAmelCase = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API UpperCAmelCase = 0 UpperCAmelCase = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": UpperCAmelCase = sigma_hat if self.state_in_first_order else sigma_next UpperCAmelCase = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": UpperCAmelCase = sigma_hat if self.state_in_first_order else sigma_next UpperCAmelCase = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": UpperCAmelCase = model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`''' ) if self.config.clip_sample: UpperCAmelCase = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order UpperCAmelCase = (sample - pred_original_sample) / sigma_hat # 3. delta timestep UpperCAmelCase = sigma_next - sigma_hat # store for 2nd order step UpperCAmelCase = derivative UpperCAmelCase = dt UpperCAmelCase = sample else: # 2. 2nd order / Heun's method UpperCAmelCase = (sample - pred_original_sample) / sigma_next UpperCAmelCase = (self.prev_derivative + derivative) / 2 # 3. take prev timestep & sample UpperCAmelCase = self.dt UpperCAmelCase = self.sample # free dt and derivative # Note, this puts the scheduler in "first order mode" UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(snake_case__ ): # mps does not support float64 UpperCAmelCase = self.timesteps.to(original_samples.device , dtype=torch.floataa ) UpperCAmelCase = timesteps.to(original_samples.device , dtype=torch.floataa ) else: UpperCAmelCase = self.timesteps.to(original_samples.device ) UpperCAmelCase = timesteps.to(original_samples.device ) UpperCAmelCase = [self.index_for_timestep(snake_case__ , snake_case__ ) for t in timesteps] UpperCAmelCase = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): UpperCAmelCase = sigma.unsqueeze(-1 ) UpperCAmelCase = original_samples + noise * sigma return noisy_samples def __len__( self ) -> int: """simple docstring""" return self.config.num_train_timesteps
673
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
673
1
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = False ): '''simple docstring''' if radian_mode: return [magnitude * cos(lowerCAmelCase ), magnitude * sin(lowerCAmelCase )] return [magnitude * cos(radians(lowerCAmelCase ) ), magnitude * sin(radians(lowerCAmelCase ) )] def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = 10**-1 ): '''simple docstring''' UpperCAmelCase = cross(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase = sum(lowerCAmelCase ) return abs(lowerCAmelCase ) < eps if __name__ == "__main__": # Test to check if it works lowerCAmelCase_ : List[Any] = array( [ polar_force(718.4, 1_8_0 - 3_0), polar_force(879.54, 4_5), polar_force(1_0_0, -9_0), ] ) lowerCAmelCase_ : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg lowerCAmelCase_ : int = array( [ polar_force(3_0 * 9.81, 1_5), polar_force(2_1_5, 1_8_0 - 4_5), polar_force(2_6_4, 9_0 - 3_0), ] ) lowerCAmelCase_ : Optional[Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg lowerCAmelCase_ : List[str] = array([[0, -2_0_0_0], [0, -1_2_0_0], [0, 1_5_6_0_0], [0, -1_2_4_0_0]]) lowerCAmelCase_ : str = array([[0, 0], [6, 0], [1_0, 0], [1_2, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
673
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = VideoToVideoSDPipeline _A : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} _A : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} _A : int = PipelineTesterMixin.required_optional_params - {'latents'} _A : List[str] = False # No `output_type`. _A : Any = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) UpperCAmelCase = CLIPTextModel(snake_case__ ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase_ ( self , snake_case__ , snake_case__=0 ) -> List[str]: """simple docstring""" UpperCAmelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(snake_case__ ) else: UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = VideoToVideoSDPipeline(**snake_case__ ) UpperCAmelCase = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) UpperCAmelCase = """np""" UpperCAmelCase = sd_pipe(**snake_case__ ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case__ ) UpperCAmelCase = video.to("""cuda""" ) UpperCAmelCase = """Spiderman is surfing""" UpperCAmelCase = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type="""pt""" ).frames UpperCAmelCase = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
673
1
"""simple docstring""" import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : Optional[Any] = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def UpperCamelCase_ ( self , snake_case__=0 ) -> str: """simple docstring""" UpperCAmelCase = np.random.RandomState(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.65_072, 0.58_492, 0.48_219, 0.55_521, 0.53_180, 0.55_939, 0.50_697, 0.39_800, 0.46_455] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) UpperCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.65_863, 0.59_425, 0.49_326, 0.56_313, 0.53_875, 0.56_627, 0.51_065, 0.39_777, 0.46_330] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) UpperCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.53_755, 0.60_786, 0.47_402, 0.49_488, 0.51_869, 0.49_819, 0.47_985, 0.38_957, 0.44_279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) UpperCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.53_755, 0.60_786, 0.47_402, 0.49_488, 0.51_869, 0.49_819, 0.47_985, 0.38_957, 0.44_279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) UpperCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.53_817, 0.60_812, 0.47_384, 0.49_530, 0.51_894, 0.49_814, 0.47_984, 0.38_958, 0.44_271] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = pipe(**snake_case__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) UpperCAmelCase = np.array([0.53_895, 0.60_808, 0.47_933, 0.49_608, 0.51_886, 0.49_950, 0.48_053, 0.38_957, 0.44_200] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = 3 * [inputs["""prompt"""]] # forward UpperCAmelCase = pipe(**snake_case__ ) UpperCAmelCase = output.images[0, -3:, -3:, -1] UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = 3 * [inputs.pop("""prompt""" )] UpperCAmelCase = pipe.tokenizer( snake_case__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=snake_case__ , return_tensors="""np""" , ) UpperCAmelCase = text_inputs["""input_ids"""] UpperCAmelCase = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] UpperCAmelCase = prompt_embeds # forward UpperCAmelCase = pipe(**snake_case__ ) UpperCAmelCase = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = 3 * ["""this is a negative prompt"""] UpperCAmelCase = negative_prompt UpperCAmelCase = 3 * [inputs["""prompt"""]] # forward UpperCAmelCase = pipe(**snake_case__ ) UpperCAmelCase = output.images[0, -3:, -3:, -1] UpperCAmelCase = self.get_dummy_inputs() UpperCAmelCase = 3 * [inputs.pop("""prompt""" )] UpperCAmelCase = [] for p in [prompt, negative_prompt]: UpperCAmelCase = pipe.tokenizer( snake_case__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=snake_case__ , return_tensors="""np""" , ) UpperCAmelCase = text_inputs["""input_ids"""] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) UpperCAmelCase , UpperCAmelCase = embeds # forward UpperCAmelCase = pipe(**snake_case__ ) UpperCAmelCase = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @nightly @require_onnxruntime @require_torch_gpu class UpperCamelCase_ ( unittest.TestCase ): @property def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = ort.SessionOptions() UpperCAmelCase = False return options def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = """A painting of a squirrel eating a burger""" np.random.seed(0 ) UpperCAmelCase = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type="""np""" ) UpperCAmelCase = output.images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCAmelCase = np.array([0.0_452, 0.0_390, 0.0_087, 0.0_350, 0.0_617, 0.0_364, 0.0_544, 0.0_523, 0.0_720] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = DDIMScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=snake_case__ , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = """open neural network exchange""" UpperCAmelCase = np.random.RandomState(0 ) UpperCAmelCase = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case__ , output_type="""np""" ) UpperCAmelCase = output.images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCAmelCase = np.array([0.2_867, 0.1_974, 0.1_481, 0.7_294, 0.7_251, 0.6_667, 0.4_194, 0.5_642, 0.6_486] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=snake_case__ , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = """open neural network exchange""" UpperCAmelCase = np.random.RandomState(0 ) UpperCAmelCase = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case__ , output_type="""np""" ) UpperCAmelCase = output.images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCAmelCase = np.array([0.2_306, 0.1_959, 0.1_593, 0.6_549, 0.6_394, 0.5_408, 0.5_065, 0.6_010, 0.6_161] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = 0 def test_callback_fn(snake_case__ , snake_case__ , snake_case__ ) -> None: UpperCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) UpperCAmelCase = latents[0, -3:, -3:, -1] UpperCAmelCase = np.array( [-0.6_772, -0.3_835, -1.2_456, 0.1_905, -1.0_974, 0.6_967, -1.9_353, 0.0_178, 1.0_167] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) UpperCAmelCase = latents[0, -3:, -3:, -1] UpperCAmelCase = np.array( [-0.3_351, 0.2_241, -0.1_837, -0.2_325, -0.6_577, 0.3_393, -0.0_241, 0.5_899, 1.3_875] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 UpperCAmelCase = False UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = """Andromeda galaxy in a bottle""" UpperCAmelCase = np.random.RandomState(0 ) pipe( prompt=snake_case__ , num_inference_steps=5 , guidance_scale=7.5 , generator=snake_case__ , callback=snake_case__ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=snake_case__ , feature_extractor=snake_case__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(snake_case__ , snake_case__ ) assert pipe.safety_checker is None UpperCAmelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(snake_case__ ) UpperCAmelCase = OnnxStableDiffusionPipeline.from_pretrained(snake_case__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None UpperCAmelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None
673
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class UpperCamelCase_ ( a_ ): _A : int = 'wav2vec2' def __init__( self , snake_case__=32 , snake_case__=7_68 , snake_case__=12 , snake_case__=12 , snake_case__=30_72 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=1e-5 , snake_case__="group" , snake_case__="gelu" , snake_case__=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case__=(5, 2, 2, 2, 2, 2, 2) , snake_case__=(10, 3, 3, 3, 3, 2, 2) , snake_case__=False , snake_case__=1_28 , snake_case__=16 , snake_case__=False , snake_case__=True , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__=3_20 , snake_case__=2 , snake_case__=0.1 , snake_case__=1_00 , snake_case__=2_56 , snake_case__=2_56 , snake_case__=0.1 , snake_case__="sum" , snake_case__=False , snake_case__=False , snake_case__=2_56 , snake_case__=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case__=(5, 3, 3, 1, 1) , snake_case__=(1, 2, 3, 1, 1) , snake_case__=5_12 , snake_case__=0 , snake_case__=1 , snake_case__=2 , snake_case__=False , snake_case__=3 , snake_case__=2 , snake_case__=3 , snake_case__=None , snake_case__=None , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) UpperCAmelCase = hidden_size UpperCAmelCase = feat_extract_norm UpperCAmelCase = feat_extract_activation UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = conv_bias UpperCAmelCase = num_conv_pos_embeddings UpperCAmelCase = num_conv_pos_embedding_groups UpperCAmelCase = len(self.conv_dim ) UpperCAmelCase = num_hidden_layers UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = feat_proj_dropout UpperCAmelCase = final_dropout UpperCAmelCase = layerdrop UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = vocab_size UpperCAmelCase = do_stable_layer_norm UpperCAmelCase = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase = apply_spec_augment UpperCAmelCase = mask_time_prob UpperCAmelCase = mask_time_length UpperCAmelCase = mask_time_min_masks UpperCAmelCase = mask_feature_prob UpperCAmelCase = mask_feature_length UpperCAmelCase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCAmelCase = num_codevectors_per_group UpperCAmelCase = num_codevector_groups UpperCAmelCase = contrastive_logits_temperature UpperCAmelCase = feat_quantizer_dropout UpperCAmelCase = num_negatives UpperCAmelCase = codevector_dim UpperCAmelCase = proj_codevector_dim UpperCAmelCase = diversity_loss_weight # ctc loss UpperCAmelCase = ctc_loss_reduction UpperCAmelCase = ctc_zero_infinity # adapter UpperCAmelCase = add_adapter UpperCAmelCase = adapter_kernel_size UpperCAmelCase = adapter_stride UpperCAmelCase = num_adapter_layers UpperCAmelCase = output_hidden_size or hidden_size UpperCAmelCase = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = xvector_output_dim @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
673
1
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class UpperCamelCase_ ( unittest.TestCase ): _A : Optional[int] = MODEL_FOR_CAUSAL_LM_MAPPING _A : Optional[Any] = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""pt""" ) # Using `do_sample=False` to force deterministic output UpperCAmelCase = text_generator("""This is a test""" , do_sample=snake_case__ ) self.assertEqual( snake_case__ , [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ] , ) UpperCAmelCase = text_generator(["""This is a test""", """This is a second test"""] ) self.assertEqual( snake_case__ , [ [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ], [ { """generated_text""": ( """This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy""" """ oscope. oscope. FiliFili@@""" ) } ], ] , ) UpperCAmelCase = text_generator("""This is a test""" , do_sample=snake_case__ , num_return_sequences=2 , return_tensors=snake_case__ ) self.assertEqual( snake_case__ , [ {"""generated_token_ids""": ANY(snake_case__ )}, {"""generated_token_ids""": ANY(snake_case__ )}, ] , ) UpperCAmelCase = text_generator.model.config.eos_token_id UpperCAmelCase = """<pad>""" UpperCAmelCase = text_generator( ["""This is a test""", """This is a second test"""] , do_sample=snake_case__ , num_return_sequences=2 , batch_size=2 , return_tensors=snake_case__ , ) self.assertEqual( snake_case__ , [ [ {"""generated_token_ids""": ANY(snake_case__ )}, {"""generated_token_ids""": ANY(snake_case__ )}, ], [ {"""generated_token_ids""": ANY(snake_case__ )}, {"""generated_token_ids""": ANY(snake_case__ )}, ], ] , ) @require_tf def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""tf""" ) # Using `do_sample=False` to force deterministic output UpperCAmelCase = text_generator("""This is a test""" , do_sample=snake_case__ ) self.assertEqual( snake_case__ , [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ] , ) UpperCAmelCase = text_generator(["""This is a test""", """This is a second test"""] , do_sample=snake_case__ ) self.assertEqual( snake_case__ , [ [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ], [ { """generated_text""": ( """This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes""" """ Cannes 閲閲Cannes Cannes Cannes 攵 please,""" ) } ], ] , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> List[str]: """simple docstring""" UpperCAmelCase = TextGenerationPipeline(model=snake_case__ , tokenizer=snake_case__ ) return text_generator, ["This is a test", "Another test"] def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = """Hello I believe in""" UpperCAmelCase = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) UpperCAmelCase = text_generator(snake_case__ ) self.assertEqual( snake_case__ , [{"""generated_text""": """Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe"""}] , ) UpperCAmelCase = text_generator(snake_case__ , stop_sequence=""" fe""" ) self.assertEqual(snake_case__ , [{"""generated_text""": """Hello I believe in fe"""}] ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = text_generator.model UpperCAmelCase = text_generator.tokenizer UpperCAmelCase = text_generator("""This is a test""" ) self.assertEqual(snake_case__ , [{"""generated_text""": ANY(snake_case__ )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) UpperCAmelCase = text_generator("""This is a test""" , return_full_text=snake_case__ ) self.assertEqual(snake_case__ , [{"""generated_text""": ANY(snake_case__ )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) UpperCAmelCase = pipeline(task="""text-generation""" , model=snake_case__ , tokenizer=snake_case__ , return_full_text=snake_case__ ) UpperCAmelCase = text_generator("""This is a test""" ) self.assertEqual(snake_case__ , [{"""generated_text""": ANY(snake_case__ )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) UpperCAmelCase = text_generator("""This is a test""" , return_full_text=snake_case__ ) self.assertEqual(snake_case__ , [{"""generated_text""": ANY(snake_case__ )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) UpperCAmelCase = text_generator(["""This is great !""", """Something else"""] , num_return_sequences=2 , do_sample=snake_case__ ) self.assertEqual( snake_case__ , [ [{"""generated_text""": ANY(snake_case__ )}, {"""generated_text""": ANY(snake_case__ )}], [{"""generated_text""": ANY(snake_case__ )}, {"""generated_text""": ANY(snake_case__ )}], ] , ) if text_generator.tokenizer.pad_token is not None: UpperCAmelCase = text_generator( ["""This is great !""", """Something else"""] , num_return_sequences=2 , batch_size=2 , do_sample=snake_case__ ) self.assertEqual( snake_case__ , [ [{"""generated_text""": ANY(snake_case__ )}, {"""generated_text""": ANY(snake_case__ )}], [{"""generated_text""": ANY(snake_case__ )}, {"""generated_text""": ANY(snake_case__ )}], ] , ) with self.assertRaises(snake_case__ ): UpperCAmelCase = text_generator("""test""" , return_full_text=snake_case__ , return_text=snake_case__ ) with self.assertRaises(snake_case__ ): UpperCAmelCase = text_generator("""test""" , return_full_text=snake_case__ , return_tensors=snake_case__ ) with self.assertRaises(snake_case__ ): UpperCAmelCase = text_generator("""test""" , return_text=snake_case__ , return_tensors=snake_case__ ) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): UpperCAmelCase = text_generator("""""" ) self.assertEqual(snake_case__ , [{"""generated_text""": ANY(snake_case__ )}] ) else: with self.assertRaises((ValueError, AssertionError) ): UpperCAmelCase = text_generator("""""" ) if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. UpperCAmelCase = ["""RwkvForCausalLM""", """XGLMForCausalLM""", """GPTNeoXForCausalLM"""] if ( tokenizer.model_max_length < 1_00_00 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError) ): text_generator("""This is a test""" * 5_00 , max_new_tokens=20 ) UpperCAmelCase = text_generator("""This is a test""" * 5_00 , handle_long_generation="""hole""" , max_new_tokens=20 ) # Hole strategy cannot work with self.assertRaises(snake_case__ ): text_generator( """This is a test""" * 5_00 , handle_long_generation="""hole""" , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" import torch # Classic `model_kwargs` UpperCAmelCase = pipeline( model="""hf-internal-testing/tiny-random-bloom""" , model_kwargs={"""device_map""": """auto""", """torch_dtype""": torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) UpperCAmelCase = pipe("""This is a test""" ) self.assertEqual( snake_case__ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) UpperCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.bfloataa ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) UpperCAmelCase = pipe("""This is a test""" ) self.assertEqual( snake_case__ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 UpperCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa ) UpperCAmelCase = pipe("""This is a test""" ) self.assertEqual( snake_case__ , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) @require_torch @require_torch_gpu def UpperCamelCase_ ( self ) -> str: """simple docstring""" import torch UpperCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device=0 , torch_dtype=torch.floataa ) pipe("""This is a test""" ) @require_torch @require_accelerate @require_torch_gpu def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" import torch UpperCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.floataa ) pipe("""This is a test""" , do_sample=snake_case__ , top_p=0.5 ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = """Hello world""" UpperCAmelCase = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) if text_generator.model.framework == "tf": UpperCAmelCase = logging.get_logger("""transformers.generation.tf_utils""" ) else: UpperCAmelCase = logging.get_logger("""transformers.generation.utils""" ) UpperCAmelCase = """Both `max_new_tokens`""" # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(snake_case__ ) as cl: UpperCAmelCase = text_generator(snake_case__ , max_length=10 , max_new_tokens=1 ) self.assertIn(snake_case__ , cl.out ) # The user only sets one -> no warning with CaptureLogger(snake_case__ ) as cl: UpperCAmelCase = text_generator(snake_case__ , max_new_tokens=1 ) self.assertNotIn(snake_case__ , cl.out ) with CaptureLogger(snake_case__ ) as cl: UpperCAmelCase = text_generator(snake_case__ , max_length=10 ) self.assertNotIn(snake_case__ , cl.out )
673
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCAmelCase_ : Optional[Any] = NewType('''DataClass''', Any) lowerCAmelCase_ : Any = NewType('''DataClassType''', Any) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = {str(lowerCAmelCase ): choice for choice in choices} return lambda lowerCAmelCase : str_to_choice.get(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( *, lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = None , **lowerCAmelCase , ): '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCAmelCase = {} if aliases is not None: UpperCAmelCase = aliases if help is not None: UpperCAmelCase = help return dataclasses.field(metadata=lowerCAmelCase , default=lowerCAmelCase , default_factory=lowerCAmelCase , **lowerCAmelCase ) class UpperCamelCase_ ( a_ ): _A : Iterable[DataClassType] def __init__( self , snake_case__ , **snake_case__ ) -> List[str]: """simple docstring""" if "formatter_class" not in kwargs: UpperCAmelCase = ArgumentDefaultsHelpFormatter super().__init__(**snake_case__ ) if dataclasses.is_dataclass(snake_case__ ): UpperCAmelCase = [dataclass_types] UpperCAmelCase = list(snake_case__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(snake_case__ ) @staticmethod def UpperCamelCase_ ( snake_case__ , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = f'''--{field.name}''' UpperCAmelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , snake_case__ ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) UpperCAmelCase = kwargs.pop("""aliases""" , [] ) if isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [aliases] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(snake_case__ , """UnionType""" ) and isinstance(snake_case__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(snake_case__ ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" f''' Problem encountered in field \'{field.name}\'.''' ) if type(snake_case__ ) not in field.type.__args__: # filter `str` in Union UpperCAmelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCAmelCase = ( field.type.__args__[0] if isinstance(snake_case__ , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCAmelCase = {} if origin_type is Literal or (isinstance(field.type , snake_case__ ) and issubclass(field.type , snake_case__ )): if origin_type is Literal: UpperCAmelCase = field.type.__args__ else: UpperCAmelCase = [x.value for x in field.type] UpperCAmelCase = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default else: UpperCAmelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCAmelCase = copy(snake_case__ ) # Hack because type=bool in argparse does not behave as we want. UpperCAmelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCAmelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCAmelCase = default # This tells argparse we accept 0 or 1 value after --field_name UpperCAmelCase = """?""" # This is the value that will get picked if we do --field_name (without value) UpperCAmelCase = True elif isclass(snake_case__ ) and issubclass(snake_case__ , snake_case__ ): UpperCAmelCase = field.type.__args__[0] UpperCAmelCase = """+""" if field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() elif field.default is dataclasses.MISSING: UpperCAmelCase = True else: UpperCAmelCase = field.type if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default elif field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() else: UpperCAmelCase = True parser.add_argument(snake_case__ , *snake_case__ , **snake_case__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCAmelCase = False parser.add_argument(f'''--no_{field.name}''' , action="""store_false""" , dest=field.name , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Any: """simple docstring""" if hasattr(snake_case__ , """_argument_group_name""" ): UpperCAmelCase = self.add_argument_group(dtype._argument_group_name ) else: UpperCAmelCase = self try: UpperCAmelCase = get_type_hints(snake_case__ ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(snake_case__ ): UpperCAmelCase = """.""".join(map(snake_case__ , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(snake_case__ ): if not field.init: continue UpperCAmelCase = type_hints[field.name] self._parse_dataclass_field(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self , snake_case__=None , snake_case__=False , snake_case__=True , snake_case__=None , snake_case__=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCAmelCase = [] if args_filename: args_files.append(Path(snake_case__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCAmelCase = ArgumentParser() args_file_parser.add_argument(snake_case__ , type=snake_case__ , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCAmelCase , UpperCAmelCase = args_file_parser.parse_known_args(args=snake_case__ ) UpperCAmelCase = vars(snake_case__ ).get(args_file_flag.lstrip("""-""" ) , snake_case__ ) if cmd_args_file_paths: args_files.extend([Path(snake_case__ ) for p in cmd_args_file_paths] ) UpperCAmelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCAmelCase = file_args + args if args is not None else file_args + sys.argv[1:] UpperCAmelCase , UpperCAmelCase = self.parse_known_args(args=snake_case__ ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in vars(snake_case__ ).items() if k in keys} for k in keys: delattr(snake_case__ , snake_case__ ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(snake_case__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = set(args.keys() ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(snake_case__ )}''' ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(snake_case__ ) , encoding="""utf-8""" ) as open_json_file: UpperCAmelCase = json.loads(open_json_file.read() ) UpperCAmelCase = self.parse_dict(snake_case__ , allow_extra_keys=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = self.parse_dict(yaml.safe_load(Path(snake_case__ ).read_text() ) , allow_extra_keys=snake_case__ ) return tuple(snake_case__ )
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if number < 0: raise ValueError("""number must not be negative""" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
673
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCAmelCase_ : List[str] = False class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self , snake_case__=32 ) -> Optional[Any]: """simple docstring""" set_seed(0 ) UpperCAmelCase = UNetaDModel(sample_size=snake_case__ , in_channels=3 , out_channels=3 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) UpperCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randint(0 , 10_00 , (4,) ).long().to(snake_case__ ) for _ in range(4 )] # train with a DDPM scheduler UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) )
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = len(lowerCAmelCase ) UpperCAmelCase = len(matrix[0] ) UpperCAmelCase = min(lowerCAmelCase , lowerCAmelCase ) for row in range(lowerCAmelCase ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , lowerCAmelCase ): UpperCAmelCase = matrix[col][row] / matrix[row][row] for i in range(lowerCAmelCase , lowerCAmelCase ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows UpperCAmelCase = True for i in range(row + 1 , lowerCAmelCase ): if matrix[i][row] != 0: UpperCAmelCase , UpperCAmelCase = matrix[i], matrix[row] UpperCAmelCase = False break if reduce: rank -= 1 for i in range(lowerCAmelCase ): UpperCAmelCase = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
673
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class UpperCamelCase_ : def __init__( self , snake_case__=2 , snake_case__=3 , snake_case__=64 , snake_case__=None ) -> List[str]: """simple docstring""" UpperCAmelCase = np.random.default_rng(snake_case__ ) UpperCAmelCase = length UpperCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) UpperCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> int: """simple docstring""" return self.length def __getitem__( self , snake_case__ ) -> Tuple: """simple docstring""" return {"x": self.x[i], "y": self.y[i]} class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[str]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> List[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a[0] + self.b[0] class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> Optional[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a + self.b def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = 16 ): '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer UpperCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) UpperCAmelCase = {"""train""": """tests/test_samples/MRPC/train.csv""", """validation""": """tests/test_samples/MRPC/dev.csv"""} UpperCAmelCase = load_dataset("""csv""" , data_files=lowerCAmelCase ) UpperCAmelCase = datasets["""train"""].unique("""label""" ) UpperCAmelCase = {v: i for i, v in enumerate(lowerCAmelCase )} def tokenize_function(lowerCAmelCase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase = tokenizer( examples["""sentence1"""] , examples["""sentence2"""] , truncation=lowerCAmelCase , max_length=lowerCAmelCase , padding="""max_length""" ) if "label" in examples: UpperCAmelCase = [label_to_id[l] for l in examples["""label"""]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase = datasets.map( lowerCAmelCase , batched=lowerCAmelCase , remove_columns=["""sentence1""", """sentence2""", """label"""] , ) def collate_fn(lowerCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return tokenizer.pad(lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. UpperCAmelCase = DataLoader(tokenized_datasets["""train"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=2 ) UpperCAmelCase = DataLoader(tokenized_datasets["""validation"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
673
1
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _lowerCAmelCase ( lowerCAmelCase = 10001 ): '''simple docstring''' try: UpperCAmelCase = int(lowerCAmelCase ) except (TypeError, ValueError): raise TypeError("""Parameter nth must be int or castable to int.""" ) from None if nth <= 0: raise ValueError("""Parameter nth must be greater than or equal to one.""" ) UpperCAmelCase = [] UpperCAmelCase = 2 while len(lowerCAmelCase ) < nth: if is_prime(lowerCAmelCase ): primes.append(lowerCAmelCase ) num += 1 else: num += 1 return primes[len(lowerCAmelCase ) - 1] if __name__ == "__main__": print(F'{solution() = }')
673
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = hidden_states.shape UpperCAmelCase = jax.image.resize( snake_case__ , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Any: """simple docstring""" UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : int = None _A : float = 0.0 _A : bool = None _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.in_channels if self.out_channels is None else self.out_channels UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = nn.Dense(snake_case__ , dtype=self.dtype ) UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Dropout(self.dropout_prob ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut UpperCAmelCase = None if use_nin_shortcut: UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self , snake_case__ , snake_case__ , snake_case__=True ) -> List[Any]: """simple docstring""" UpperCAmelCase = hidden_states UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) UpperCAmelCase = self.time_emb_proj(nn.swish(snake_case__ ) ) UpperCAmelCase = jnp.expand_dims(jnp.expand_dims(snake_case__ , 1 ) , 1 ) UpperCAmelCase = hidden_states + temb UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.dropout(snake_case__ , snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) if self.conv_shortcut is not None: UpperCAmelCase = self.conv_shortcut(snake_case__ ) return hidden_states + residual
673
1
"""simple docstring""" import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=5_12 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=4 , ) -> Optional[int]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_attention_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_choices def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_attention_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = FlaxAlbertModelTester(self ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase = model_class_name.from_pretrained("""albert-base-v2""" ) UpperCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case__ ) @require_flax class UpperCamelCase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = FlaxAlbertModel.from_pretrained("""albert-base-v2""" ) UpperCAmelCase = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) UpperCAmelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ )[0] UpperCAmelCase = (1, 11, 7_68) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , snake_case__ , atol=1e-4 ) )
673
"""simple docstring""" from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=30 , snake_case__=2 , snake_case__=3 , snake_case__=True , snake_case__=True , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=10 , snake_case__=0.02 , snake_case__=3 , snake_case__=None , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 1 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModel(config=snake_case__ ) UpperCAmelCase = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) UpperCAmelCase = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[int] = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _A : Optional[Any] = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) _A : Optional[int] = False _A : Any = False _A : List[str] = False def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(snake_case__ ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=snake_case__ , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**snake_case__ ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case__ ) UpperCAmelCase = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , snake_case__ , atol=1e-4 )
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=5_12 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = NystromformerForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = NystromformerForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = self.num_choices UpperCAmelCase = NystromformerForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[Any] = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) _A : Optional[Any] = ( { 'feature-extraction': NystromformerModel, 'fill-mask': NystromformerForMaskedLM, 'question-answering': NystromformerForQuestionAnswering, 'text-classification': NystromformerForSequenceClassification, 'token-classification': NystromformerForTokenClassification, 'zero-shot': NystromformerForSequenceClassification, } if is_torch_available() else {} ) _A : int = False _A : Dict = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = NystromformerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case__ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = NystromformerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): UpperCAmelCase = model(snake_case__ )[0] UpperCAmelCase = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """the [MASK] of Belgium is Brussels""" UpperCAmelCase = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = tokenizer(snake_case__ , return_tensors="""pt""" ) with torch.no_grad(): UpperCAmelCase = model(encoding.input_ids ).logits UpperCAmelCase = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(snake_case__ ) , """capital""" )
673
1
"""simple docstring""" import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer lowerCAmelCase_ : Dict = logging.get_logger(__name__) lowerCAmelCase_ : Dict = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase_ : Tuple = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } lowerCAmelCase_ : List[Any] = { '''allenai/led-base-16384''': 1_6_3_8_4, } class UpperCamelCase_ ( a_ ): _A : List[str] = VOCAB_FILES_NAMES _A : Any = PRETRAINED_VOCAB_FILES_MAP _A : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : int = LEDTokenizer _A : Dict = ['input_ids', 'attention_mask'] def __init__( self , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__="replace" , snake_case__="<s>" , snake_case__="</s>" , snake_case__="</s>" , snake_case__="<s>" , snake_case__="<unk>" , snake_case__="<pad>" , snake_case__="<mask>" , snake_case__=False , snake_case__=True , **snake_case__ , ) -> Dict: """simple docstring""" super().__init__( snake_case__ , snake_case__ , tokenizer_file=snake_case__ , errors=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ , **snake_case__ , ) UpperCAmelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , snake_case__ ) != add_prefix_space: UpperCAmelCase = getattr(snake_case__ , pre_tok_state.pop("""type""" ) ) UpperCAmelCase = add_prefix_space UpperCAmelCase = pre_tok_class(**snake_case__ ) UpperCAmelCase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` UpperCAmelCase = """post_processor""" UpperCAmelCase = getattr(self.backend_tokenizer , snake_case__ , snake_case__ ) if tokenizer_component_instance: UpperCAmelCase = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: UpperCAmelCase = tuple(state["""sep"""] ) if "cls" in state: UpperCAmelCase = tuple(state["""cls"""] ) UpperCAmelCase = False if state.get("""add_prefix_space""" , snake_case__ ) != add_prefix_space: UpperCAmelCase = add_prefix_space UpperCAmelCase = True if state.get("""trim_offsets""" , snake_case__ ) != trim_offsets: UpperCAmelCase = trim_offsets UpperCAmelCase = True if changes_to_apply: UpperCAmelCase = getattr(snake_case__ , state.pop("""type""" ) ) UpperCAmelCase = component_class(**snake_case__ ) setattr(self.backend_tokenizer , snake_case__ , snake_case__ ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def UpperCamelCase_ ( self ) -> str: """simple docstring""" if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""" ) return None return str(self._mask_token ) @mask_token.setter def UpperCamelCase_ ( self , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else value UpperCAmelCase = value def UpperCamelCase_ ( self , *snake_case__ , **snake_case__ ) -> BatchEncoding: """simple docstring""" UpperCAmelCase = kwargs.get("""is_split_into_words""" , snake_case__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' """to use it with pretokenized inputs.""" ) return super()._batch_encode_plus(*snake_case__ , **snake_case__ ) def UpperCamelCase_ ( self , *snake_case__ , **snake_case__ ) -> BatchEncoding: """simple docstring""" UpperCAmelCase = kwargs.get("""is_split_into_words""" , snake_case__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' """to use it with pretokenized inputs.""" ) return super()._encode_plus(*snake_case__ , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase = self._tokenizer.model.save(snake_case__ , name=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__=None ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None , snake_case__ = PaddingStrategy.DO_NOT_PAD , snake_case__ = None , snake_case__ = None , ) -> dict: """simple docstring""" UpperCAmelCase = super()._pad( encoded_inputs=snake_case__ , max_length=snake_case__ , padding_strategy=snake_case__ , pad_to_multiple_of=snake_case__ , return_attention_mask=snake_case__ , ) # Load from model defaults if return_attention_mask is None: UpperCAmelCase = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: UpperCAmelCase = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. UpperCAmelCase = len(encoded_inputs["""global_attention_mask"""] ) != len(snake_case__ ) if needs_to_be_padded: UpperCAmelCase = len(snake_case__ ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` UpperCAmelCase = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": UpperCAmelCase = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
673
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Optional[int] = False def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return TrainCommand(lowerCAmelCase ) class UpperCamelCase_ ( a_ ): @staticmethod def UpperCamelCase_ ( snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = parser.add_parser("""train""" , help="""CLI tool to train a model on a task.""" ) train_parser.add_argument( """--train_data""" , type=snake_case__ , required=snake_case__ , help="""path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.""" , ) train_parser.add_argument( """--column_label""" , type=snake_case__ , default=0 , help="""Column of the dataset csv file with example labels.""" ) train_parser.add_argument( """--column_text""" , type=snake_case__ , default=1 , help="""Column of the dataset csv file with example texts.""" ) train_parser.add_argument( """--column_id""" , type=snake_case__ , default=2 , help="""Column of the dataset csv file with example ids.""" ) train_parser.add_argument( """--skip_first_row""" , action="""store_true""" , help="""Skip the first row of the csv file (headers).""" ) train_parser.add_argument("""--validation_data""" , type=snake_case__ , default="""""" , help="""path to validation dataset.""" ) train_parser.add_argument( """--validation_split""" , type=snake_case__ , default=0.1 , help="""if validation dataset is not provided, fraction of train dataset to use as validation dataset.""" , ) train_parser.add_argument("""--output""" , type=snake_case__ , default="""./""" , help="""path to saved the trained model.""" ) train_parser.add_argument( """--task""" , type=snake_case__ , default="""text_classification""" , help="""Task to train the model on.""" ) train_parser.add_argument( """--model""" , type=snake_case__ , default="""bert-base-uncased""" , help="""Model's name or path to stored model.""" ) train_parser.add_argument("""--train_batch_size""" , type=snake_case__ , default=32 , help="""Batch size for training.""" ) train_parser.add_argument("""--valid_batch_size""" , type=snake_case__ , default=64 , help="""Batch size for validation.""" ) train_parser.add_argument("""--learning_rate""" , type=snake_case__ , default=3e-5 , help="""Learning rate.""" ) train_parser.add_argument("""--adam_epsilon""" , type=snake_case__ , default=1e-08 , help="""Epsilon for Adam optimizer.""" ) train_parser.set_defaults(func=snake_case__ ) def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = logging.get_logger("""transformers-cli/training""" ) UpperCAmelCase = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=snake_case__ ) UpperCAmelCase = args.output UpperCAmelCase = args.column_label UpperCAmelCase = args.column_text UpperCAmelCase = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": UpperCAmelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = args.validation_split UpperCAmelCase = args.train_batch_size UpperCAmelCase = args.valid_batch_size UpperCAmelCase = args.learning_rate UpperCAmelCase = args.adam_epsilon def UpperCamelCase_ ( self ) -> Any: """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" raise NotImplementedError def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
673
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer lowerCAmelCase_ : List[Any] = logging.get_logger(__name__) lowerCAmelCase_ : List[str] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase_ : Union[str, Any] = { '''vocab_file''': {'''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'''}, '''tokenizer_file''': { '''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json''' }, } lowerCAmelCase_ : Tuple = {'''mobilebert-uncased''': 5_1_2} lowerCAmelCase_ : List[Any] = {} class UpperCamelCase_ ( a_ ): _A : Optional[int] = VOCAB_FILES_NAMES _A : List[str] = PRETRAINED_VOCAB_FILES_MAP _A : str = PRETRAINED_INIT_CONFIGURATION _A : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : int = MobileBertTokenizer def __init__( self , snake_case__=None , snake_case__=None , snake_case__=True , snake_case__="[UNK]" , snake_case__="[SEP]" , snake_case__="[PAD]" , snake_case__="[CLS]" , snake_case__="[MASK]" , snake_case__=True , snake_case__=None , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__( snake_case__ , tokenizer_file=snake_case__ , do_lower_case=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , tokenize_chinese_chars=snake_case__ , strip_accents=snake_case__ , **snake_case__ , ) UpperCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , snake_case__ ) != do_lower_case or normalizer_state.get("""strip_accents""" , snake_case__ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , snake_case__ ) != tokenize_chinese_chars ): UpperCAmelCase = getattr(snake_case__ , normalizer_state.pop("""type""" ) ) UpperCAmelCase = do_lower_case UpperCAmelCase = strip_accents UpperCAmelCase = tokenize_chinese_chars UpperCAmelCase = normalizer_class(**snake_case__ ) UpperCAmelCase = do_lower_case def UpperCamelCase_ ( self , snake_case__ , snake_case__=None ) -> Dict: """simple docstring""" UpperCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase = self._tokenizer.model.save(snake_case__ , name=snake_case__ ) return tuple(snake_case__ )
673
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=sys.maxsize ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """bilinear""" UpperCAmelCase = max_size UpperCAmelCase = short_edge_length def __call__( self , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = [] for img in imgs: UpperCAmelCase , UpperCAmelCase = img.shape[:2] # later: provide list and randomly choose index for resize UpperCAmelCase = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img UpperCAmelCase = size * 1.0 / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size if max(snake_case__ , snake_case__ ) > self.max_size: UpperCAmelCase = self.max_size * 1.0 / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase = int(neww + 0.5 ) UpperCAmelCase = int(newh + 0.5 ) if img.dtype == np.uinta: UpperCAmelCase = Image.fromarray(snake_case__ ) UpperCAmelCase = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) UpperCAmelCase = np.asarray(snake_case__ ) else: UpperCAmelCase = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw UpperCAmelCase = nn.functional.interpolate( snake_case__ , (newh, neww) , mode=self.interp_method , align_corners=snake_case__ ).squeeze(0 ) img_augs.append(snake_case__ ) return img_augs class UpperCamelCase_ : def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) UpperCAmelCase = cfg.INPUT.FORMAT UpperCAmelCase = cfg.SIZE_DIVISIBILITY UpperCAmelCase = cfg.PAD_VALUE UpperCAmelCase = cfg.INPUT.MAX_SIZE_TEST UpperCAmelCase = cfg.MODEL.DEVICE UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = lambda snake_case__ : (x - self.pixel_mean) / self.pixel_std def UpperCamelCase_ ( self , snake_case__ ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = tuple(max(snake_case__ ) for s in zip(*[img.shape for img in images] ) ) UpperCAmelCase = [im.shape[-2:] for im in images] UpperCAmelCase = [ nn.functional.pad( snake_case__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(snake_case__ , snake_case__ ) ] return torch.stack(snake_case__ ), torch.tensor(snake_case__ ) def __call__( self , snake_case__ , snake_case__=False ) -> Optional[Any]: """simple docstring""" with torch.no_grad(): if not isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [images] if single_image: assert len(snake_case__ ) == 1 for i in range(len(snake_case__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(snake_case__ , images.pop(snake_case__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( snake_case__ , torch.as_tensor(img_tensorize(images.pop(snake_case__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge UpperCAmelCase = torch.tensor([im.shape[:2] for im in images] ) UpperCAmelCase = self.aug(snake_case__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic UpperCAmelCase = [self.normalizer(snake_case__ ) for x in images] # now pad them to do the following operations UpperCAmelCase , UpperCAmelCase = self.pad(snake_case__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad UpperCAmelCase = torch.true_divide(snake_case__ , snake_case__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' assert torch.isfinite(lowerCAmelCase ).all(), "Box tensor contains infinite or NaN!" UpperCAmelCase , UpperCAmelCase = box_size tensor[:, 0].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 1].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 2].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 3].clamp_(min=0 , max=lowerCAmelCase )
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : List[str] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase = """""" else: UpperCAmelCase = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = dct.pop(lowerCAmelCase ) UpperCAmelCase = val def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase = 1000 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = int(deit_name[-6:-4] ) UpperCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): UpperCAmelCase = 192 UpperCAmelCase = 768 UpperCAmelCase = 12 UpperCAmelCase = 3 elif deit_name[9:].startswith("""small""" ): UpperCAmelCase = 384 UpperCAmelCase = 1536 UpperCAmelCase = 12 UpperCAmelCase = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 # load original model from timm UpperCAmelCase = timm.create_model(lowerCAmelCase , pretrained=lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase = timm_model.state_dict() UpperCAmelCase = create_rename_keys(lowerCAmelCase , lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) read_in_q_k_v(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model UpperCAmelCase = DeiTForImageClassificationWithTeacher(lowerCAmelCase ).eval() model.load_state_dict(lowerCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase = DeiTImageProcessor(size=lowerCAmelCase , crop_size=config.image_size ) UpperCAmelCase = image_processor(images=prepare_img() , return_tensors="""pt""" ) UpperCAmelCase = encoding["""pixel_values"""] UpperCAmelCase = model(lowerCAmelCase ) UpperCAmelCase = timm_model(lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase_ : str = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
673
1
"""simple docstring""" import sys lowerCAmelCase_ : int = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 1 for digit in s: product *= int(lowerCAmelCase ) return product def _lowerCAmelCase ( lowerCAmelCase = N ): '''simple docstring''' UpperCAmelCase = -sys.maxsize - 1 UpperCAmelCase = n[:13] UpperCAmelCase = 13 while cur_index < len(lowerCAmelCase ) - 13: if int(n[cur_index] ) >= int(substr[0] ): UpperCAmelCase = substr[1:] + n[cur_index] cur_index += 1 else: UpperCAmelCase = max(lowerCAmelCase , str_eval(lowerCAmelCase ) ) UpperCAmelCase = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(F'{solution() = }')
673
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = 32 , snake_case__ = True , snake_case__ = 1 / 2_55 , snake_case__ = True , snake_case__ = True , snake_case__ = [0.48_145_466, 0.4_578_275, 0.40_821_073] , snake_case__ = [0.26_862_954, 0.26_130_258, 0.27_577_711] , snake_case__ = True , snake_case__=7 , snake_case__=30 , snake_case__=4_00 , snake_case__=3 , ) -> List[str]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = do_resize UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_88} UpperCAmelCase = size_divisor UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_pad UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase_ ( self , snake_case__ , snake_case__=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = image_inputs[0] if isinstance(snake_case__ , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] UpperCAmelCase = size / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size UpperCAmelCase = int((13_33 / 8_00) * size ) if max(snake_case__ , snake_case__ ) > max_size: UpperCAmelCase = max_size / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase , UpperCAmelCase = int(newh + 0.5 ), int(neww + 0.5 ) UpperCAmelCase , UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """size_divisor""" ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = int(lowerCAmelCase ) if n_element < 1: UpperCAmelCase = ValueError("""a should be a positive number""" ) raise my_error UpperCAmelCase = [1] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = (0, 0, 0) UpperCAmelCase = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": lowerCAmelCase_ : Dict = input('''Enter the last number (nth term) of the Hamming Number Series: ''') print('''Formula of Hamming Number Series => 2^i * 3^j * 5^k''') lowerCAmelCase_ : str = hamming(int(n)) print('''-----------------------------------------------------''') print(F'The list with nth numbers is: {hamming_numbers}') print('''-----------------------------------------------------''')
673
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[str] = XLMRobertaTokenizer _A : List[str] = XLMRobertaTokenizerFast _A : Optional[Any] = True _A : List[str] = True def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(snake_case__ ) , 10_02 ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(snake_case__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) UpperCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=snake_case__ ) UpperCAmelCase = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = """I was born in 92000, and this is falsé.""" UpperCAmelCase = tokenizer.tokenize(snake_case__ ) UpperCAmelCase = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
673
1
"""simple docstring""" import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def _lowerCAmelCase ( ): '''simple docstring''' raise RuntimeError("""CUDA out of memory.""" ) class UpperCamelCase_ ( nn.Module ): def __init__( self ) -> Union[str, Any]: """simple docstring""" super().__init__() UpperCAmelCase = nn.Linear(3 , 4 ) UpperCAmelCase = nn.BatchNormad(4 ) UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase_ ( self , snake_case__ ) -> Dict: """simple docstring""" return self.lineara(self.batchnorm(self.lineara(snake_case__ ) ) ) class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(snake_case__ ): nonlocal batch_sizes batch_sizes.append(snake_case__ ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(snake_case__ , [1_28, 64, 32, 16, 8] ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(snake_case__ , snake_case__ ): nonlocal batch_sizes batch_sizes.append(snake_case__ ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga UpperCAmelCase , UpperCAmelCase = mock_training_loop_function("""hello""" ) self.assertListEqual(snake_case__ , [1_28, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, """hello"""] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(snake_case__ ): pass with self.assertRaises(snake_case__ ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(snake_case__ ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(snake_case__ ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(snake_case__ , snake_case__ , snake_case__ ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(snake_case__ ) as cm: mock_training_loop_function(1_28 , """hello""" , """world""" ) self.assertIn("""Batch size was passed into `f`""" , cm.exception.args[0] ) self.assertIn("""`f(arg1='hello', arg2='world')""" , cm.exception.args[0] ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(snake_case__ ): raise ValueError("""Oops, we had an error!""" ) with self.assertRaises(snake_case__ ) as cm: mock_training_loop_function() self.assertIn("""Oops, we had an error!""" , cm.exception.args[0] ) @require_cuda def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = torch.cuda.memory_allocated() UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , snake_case__ ) UpperCAmelCase = release_memory(snake_case__ ) self.assertEqual(torch.cuda.memory_allocated() , snake_case__ )
673
"""simple docstring""" import socket def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase = socket.gethostname() UpperCAmelCase = 12312 sock.connect((host, port) ) sock.send(b"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: UpperCAmelCase = sock.recv(1024 ) if not data: break out_file.write(lowerCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
673
1
"""simple docstring""" import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class UpperCamelCase_ ( a_ ): @require_torch def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ UpperCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ UpperCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache UpperCAmelCase = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(snake_case__ ) BertModel.from_pretrained(snake_case__ ) BertTokenizer.from_pretrained(snake_case__ ) pipeline(task="""fill-mask""" , model=snake_case__ ) # baseline - just load from_pretrained with normal network UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed UpperCAmelCase = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase = """1""" UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ UpperCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ UpperCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache UpperCAmelCase = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(snake_case__ ) BertModel.from_pretrained(snake_case__ ) BertTokenizer.from_pretrained(snake_case__ ) pipeline(task="""fill-mask""" , model=snake_case__ ) # baseline - just load from_pretrained with normal network UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed UpperCAmelCase = self.get_env() UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer """ UpperCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert-sharded\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print(\"success\") """ UpperCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\") socket.socket = offline_socket """ # baseline - just load from_pretrained with normal network UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed UpperCAmelCase = self.get_env() UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # next emulate no network UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase = """1""" UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = """ from transformers import pipeline """ UpperCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" pipe = pipeline(model=mname) """ UpperCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\") socket.socket = offline_socket """ UpperCAmelCase = self.get_env() UpperCAmelCase = """1""" UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, mock, run] )] UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( """You cannot infer task automatically within `pipeline` when using offline mode""" , result.stderr.decode().replace("""\n""" , """""" ) , ) @require_torch def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """ from transformers import AutoModel """ UpperCAmelCase = """ mname = \"hf-internal-testing/test_dynamic_model\" AutoModel.from_pretrained(mname, trust_remote_code=True) print(\"success\") """ # baseline - just load from_pretrained with normal network UpperCAmelCase = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed UpperCAmelCase = self.get_env() UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase = """1""" UpperCAmelCase = subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() )
673
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
1
"""simple docstring""" import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCAmelCase_ : List[str] = get_tests_dir('''fixtures/dummy-config.json''') class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = 0 def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("""transformers.models.auto""" ) ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = AutoConfig.from_pretrained("""bert-base-uncased""" ) self.assertIsInstance(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ ) self.assertIsInstance(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ ) self.assertIsInstance(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = AutoConfig.for_model("""roberta""" ) self.assertIsInstance(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. UpperCAmelCase = os.path.join(snake_case__ , """fake-roberta""" ) os.makedirs(snake_case__ , exist_ok=snake_case__ ) with open(os.path.join(snake_case__ , """config.json""" ) , """w""" ) as f: f.write(json.dumps({} ) ) UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ ) self.assertEqual(type(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" try: AutoConfig.register("""custom""" , snake_case__ ) # Wrong model type will raise an error with self.assertRaises(snake_case__ ): AutoConfig.register("""model""" , snake_case__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(snake_case__ ): AutoConfig.register("""bert""" , snake_case__ ) # Now that the config is registered, it can be used as any other config with the auto-API UpperCAmelCase = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(snake_case__ ) UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ ) self.assertIsInstance(snake_case__ , snake_case__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" with self.assertRaisesRegex( snake_case__ , """bert-base is not a local folder and is not a valid model identifier""" ): UpperCAmelCase = AutoConfig.from_pretrained("""bert-base""" ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" with self.assertRaisesRegex( snake_case__ , R"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ , revision="""aaaaaa""" ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" with self.assertRaisesRegex( snake_case__ , """hf-internal-testing/no-config-test-repo does not appear to have a file named config.json.""" , ): UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/no-config-test-repo""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" with self.assertRaises(snake_case__ ): UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(snake_case__ ): UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=snake_case__ ) UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=snake_case__ ) self.assertEqual(config.__class__.__name__ , """NewModelConfig""" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(snake_case__ ) UpperCAmelCase = AutoConfig.from_pretrained(snake_case__ , trust_remote_code=snake_case__ ) self.assertEqual(reloaded_config.__class__.__name__ , """NewModelConfig""" ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" class UpperCamelCase_ ( a_ ): _A : List[str] = 'new-model' try: AutoConfig.register("""new-model""" , snake_case__ ) # If remote code is not set, the default is to use local UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" ) self.assertEqual(config.__class__.__name__ , """NewModelConfigLocal""" ) # If remote code is disabled, we load the local one. UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=snake_case__ ) self.assertEqual(config.__class__.__name__ , """NewModelConfigLocal""" ) # If remote is enabled, we load from the Hub UpperCAmelCase = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=snake_case__ ) self.assertEqual(config.__class__.__name__ , """NewModelConfig""" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
673
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = list(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): UpperCAmelCase = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _lowerCAmelCase ( lowerCAmelCase = None , lowerCAmelCase = 128 ): '''simple docstring''' if function is None: return functools.partial(lowerCAmelCase , starting_batch_size=lowerCAmelCase ) UpperCAmelCase = starting_batch_size def decorator(*lowerCAmelCase , **lowerCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() UpperCAmelCase = list(inspect.signature(lowerCAmelCase ).parameters.keys() ) # Guard against user error if len(lowerCAmelCase ) < (len(lowerCAmelCase ) + 1): UpperCAmelCase = """, """.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) except Exception as e: if should_reduce_batch_size(lowerCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
673
1
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' with open(lowerCAmelCase , """r""" ) as fh: fcntl.flock(lowerCAmelCase , fcntl.LOCK_EX ) try: print(*lowerCAmelCase ) finally: fcntl.flock(lowerCAmelCase , fcntl.LOCK_UN ) lowerCAmelCase_ : Dict = int(os.environ['''LOCAL_RANK''']) torch.cuda.set_device(local_rank) lowerCAmelCase_ : int = torch.device('''cuda''', local_rank) lowerCAmelCase_ : List[str] = socket.gethostname() lowerCAmelCase_ : str = F'[{hostname}-{local_rank}]' try: # test distributed dist.init_process_group('''nccl''') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank lowerCAmelCase_ : Tuple = dist.get_rank() lowerCAmelCase_ : Any = dist.get_world_size() printflock(F'{gpu} is OK (global rank: {rank}/{world_size})') dist.barrier() if rank == 0: printflock(F'pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}') except Exception: printflock(F'{gpu} is broken') raise
673
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
1
"""simple docstring""" import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase_ : Dict = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = { """attention_cell""": """multi_head""", """num_layers""": 4, """units""": 1024, """hidden_size""": 768, """max_length""": 512, """num_heads""": 8, """scaled""": True, """dropout""": 0.1, """use_residual""": True, """embed_size""": 1024, """embed_dropout""": 0.1, """word_embed""": None, """layer_norm_eps""": 1e-5, """token_type_vocab_size""": 2, } UpperCAmelCase = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py UpperCAmelCase = BERTEncoder( attention_cell=predefined_args["""attention_cell"""] , num_layers=predefined_args["""num_layers"""] , units=predefined_args["""units"""] , hidden_size=predefined_args["""hidden_size"""] , max_length=predefined_args["""max_length"""] , num_heads=predefined_args["""num_heads"""] , scaled=predefined_args["""scaled"""] , dropout=predefined_args["""dropout"""] , output_attention=lowerCAmelCase , output_all_encodings=lowerCAmelCase , use_residual=predefined_args["""use_residual"""] , activation=predefined_args.get("""activation""" , """gelu""" ) , layer_norm_eps=predefined_args.get("""layer_norm_eps""" , lowerCAmelCase ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later UpperCAmelCase = """openwebtext_ccnews_stories_books_cased""" # Specify download folder to Gluonnlp's vocab UpperCAmelCase = os.path.join(get_home_dir() , """models""" ) UpperCAmelCase = _load_vocab(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , cls=lowerCAmelCase ) UpperCAmelCase = nlp.model.BERTModel( lowerCAmelCase , len(lowerCAmelCase ) , units=predefined_args["""units"""] , embed_size=predefined_args["""embed_size"""] , embed_dropout=predefined_args["""embed_dropout"""] , word_embed=predefined_args["""word_embed"""] , use_pooler=lowerCAmelCase , use_token_type_embed=lowerCAmelCase , token_type_vocab_size=predefined_args["""token_type_vocab_size"""] , use_classifier=lowerCAmelCase , use_decoder=lowerCAmelCase , ) original_bort.load_parameters(lowerCAmelCase , cast_dtype=lowerCAmelCase , ignore_extra=lowerCAmelCase ) UpperCAmelCase = original_bort._collect_params_with_prefix() # Build our config 🤗 UpperCAmelCase = { """architectures""": ["""BertForMaskedLM"""], """attention_probs_dropout_prob""": predefined_args["""dropout"""], """hidden_act""": """gelu""", """hidden_dropout_prob""": predefined_args["""dropout"""], """hidden_size""": predefined_args["""embed_size"""], """initializer_range""": 0.02, """intermediate_size""": predefined_args["""hidden_size"""], """layer_norm_eps""": predefined_args["""layer_norm_eps"""], """max_position_embeddings""": predefined_args["""max_length"""], """model_type""": """bort""", """num_attention_heads""": predefined_args["""num_heads"""], """num_hidden_layers""": predefined_args["""num_layers"""], """pad_token_id""": 1, # 2 = BERT, 1 = RoBERTa """type_vocab_size""": 1, # 2 = BERT, 1 = RoBERTa """vocab_size""": len(lowerCAmelCase ), } UpperCAmelCase = BertConfig.from_dict(lowerCAmelCase ) UpperCAmelCase = BertForMaskedLM(lowerCAmelCase ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(lowerCAmelCase ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = hf_param.shape UpperCAmelCase = to_torch(params[gluon_param] ) UpperCAmelCase = gluon_param.shape assert ( shape_hf == shape_gluon ), F'''The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers''' return gluon_param UpperCAmelCase = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , """word_embed.0.weight""" ) UpperCAmelCase = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , """encoder.position_weight""" ) UpperCAmelCase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , """encoder.layer_norm.beta""" ) UpperCAmelCase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , """encoder.layer_norm.gamma""" ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) UpperCAmelCase = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): UpperCAmelCase = hf_bort_model.bert.encoder.layer[i] # self attention UpperCAmelCase = layer.attention.self UpperCAmelCase = check_and_map_params( self_attn.key.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_key.bias''' ) UpperCAmelCase = check_and_map_params( self_attn.key.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_key.weight''' ) UpperCAmelCase = check_and_map_params( self_attn.query.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_query.bias''' ) UpperCAmelCase = check_and_map_params( self_attn.query.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_query.weight''' ) UpperCAmelCase = check_and_map_params( self_attn.value.bias.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_value.bias''' ) UpperCAmelCase = check_and_map_params( self_attn.value.weight.data , F'''encoder.transformer_cells.{i}.attention_cell.proj_value.weight''' ) # self attention output UpperCAmelCase = layer.attention.output UpperCAmelCase = check_and_map_params( self_output.dense.bias , F'''encoder.transformer_cells.{i}.proj.bias''' ) UpperCAmelCase = check_and_map_params( self_output.dense.weight , F'''encoder.transformer_cells.{i}.proj.weight''' ) UpperCAmelCase = check_and_map_params( self_output.LayerNorm.bias , F'''encoder.transformer_cells.{i}.layer_norm.beta''' ) UpperCAmelCase = check_and_map_params( self_output.LayerNorm.weight , F'''encoder.transformer_cells.{i}.layer_norm.gamma''' ) # intermediate UpperCAmelCase = layer.intermediate UpperCAmelCase = check_and_map_params( intermediate.dense.bias , F'''encoder.transformer_cells.{i}.ffn.ffn_1.bias''' ) UpperCAmelCase = check_and_map_params( intermediate.dense.weight , F'''encoder.transformer_cells.{i}.ffn.ffn_1.weight''' ) # output UpperCAmelCase = layer.output UpperCAmelCase = check_and_map_params( bert_output.dense.bias , F'''encoder.transformer_cells.{i}.ffn.ffn_2.bias''' ) UpperCAmelCase = check_and_map_params( bert_output.dense.weight , F'''encoder.transformer_cells.{i}.ffn.ffn_2.weight''' ) UpperCAmelCase = check_and_map_params( bert_output.LayerNorm.bias , F'''encoder.transformer_cells.{i}.ffn.layer_norm.beta''' ) UpperCAmelCase = check_and_map_params( bert_output.LayerNorm.weight , F'''encoder.transformer_cells.{i}.ffn.layer_norm.gamma''' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models UpperCAmelCase = RobertaTokenizer.from_pretrained("""roberta-base""" ) UpperCAmelCase = tokenizer.encode_plus(lowerCAmelCase )["""input_ids"""] # Get gluon output UpperCAmelCase = mx.nd.array([input_ids] ) UpperCAmelCase = original_bort(inputs=lowerCAmelCase , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(lowerCAmelCase ) UpperCAmelCase = BertModel.from_pretrained(lowerCAmelCase ) hf_bort_model.eval() UpperCAmelCase = tokenizer.encode_plus(lowerCAmelCase , return_tensors="""pt""" ) UpperCAmelCase = hf_bort_model(**lowerCAmelCase )[0] UpperCAmelCase = output_gluon[0].asnumpy() UpperCAmelCase = output_hf[0].detach().numpy() UpperCAmelCase = np.max(np.abs(hf_layer - gluon_layer ) ).item() UpperCAmelCase = np.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-3 ) if success: print("""✔️ Both model do output the same tensors""" ) else: print("""❌ Both model do **NOT** output the same tensors""" ) print("""Absolute difference is:""" , lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase_ : int = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
673
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ : List[Any] = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[Any] = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Tuple = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : int = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Union[str, Any] = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Union[str, Any] = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys lowerCAmelCase_ : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
673
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( a_ ): _A : Optional[int] = 'facebook/bart-large-mnli' _A : Union[str, Any] = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) _A : Dict = 'text_classifier' _A : Union[str, Any] = AutoTokenizer _A : Tuple = AutoModelForSequenceClassification _A : Optional[int] = ['text', ['text']] _A : Dict = ['text'] def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setup() UpperCAmelCase = self.model.config UpperCAmelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = labels return self.pre_processor( [text] * len(snake_case__ ) , [f'''This example is {label}''' for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def UpperCamelCase_ ( self , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = outputs.logits UpperCAmelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
673
1
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase_ ( a_ ): _A : Union[List[PIL.Image.Image], np.ndarray] _A : Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
673
1
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None ): '''simple docstring''' if version.parse(hfh.__version__ ).release < version.parse("""0.11.0""" ).release: # old versions of hfh don't url-encode the file path UpperCAmelCase = quote(lowerCAmelCase ) return hfh.hf_hub_url(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" , revision=lowerCAmelCase )
673
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : Any = { '''configuration_encodec''': [ '''ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EncodecConfig''', ], '''feature_extraction_encodec''': ['''EncodecFeatureExtractor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ '''ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EncodecModel''', '''EncodecPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys lowerCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
1
"""simple docstring""" import operator def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = False , lowerCAmelCase = None ): '''simple docstring''' UpperCAmelCase = operator.lt if reverse else operator.gt UpperCAmelCase = solution or [] if not arr: return solution UpperCAmelCase = [arr.pop(0 )] for i, item in enumerate(lowerCAmelCase ): if _operator(lowerCAmelCase , sublist[-1] ): sublist.append(lowerCAmelCase ) arr.pop(lowerCAmelCase ) # merging sublist into solution list if not solution: solution.extend(lowerCAmelCase ) else: while sublist: UpperCAmelCase = sublist.pop(0 ) for i, xx in enumerate(lowerCAmelCase ): if not _operator(lowerCAmelCase , lowerCAmelCase ): solution.insert(lowerCAmelCase , lowerCAmelCase ) break else: solution.append(lowerCAmelCase ) strand_sort(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
673
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase = 1000 ): '''simple docstring''' UpperCAmelCase = 2**power UpperCAmelCase = str(lowerCAmelCase ) UpperCAmelCase = list(lowerCAmelCase ) UpperCAmelCase = 0 for i in list_num: sum_of_num += int(lowerCAmelCase ) return sum_of_num if __name__ == "__main__": lowerCAmelCase_ : Union[str, Any] = int(input('''Enter the power of 2: ''').strip()) print('''2 ^ ''', power, ''' = ''', 2**power) lowerCAmelCase_ : Any = solution(power) print('''Sum of the digits is: ''', result)
673
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = VideoToVideoSDPipeline _A : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} _A : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} _A : int = PipelineTesterMixin.required_optional_params - {'latents'} _A : List[str] = False # No `output_type`. _A : Any = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) UpperCAmelCase = CLIPTextModel(snake_case__ ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase_ ( self , snake_case__ , snake_case__=0 ) -> List[str]: """simple docstring""" UpperCAmelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(snake_case__ ) else: UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = VideoToVideoSDPipeline(**snake_case__ ) UpperCAmelCase = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) UpperCAmelCase = """np""" UpperCAmelCase = sd_pipe(**snake_case__ ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case__ ) UpperCAmelCase = video.to("""cuda""" ) UpperCAmelCase = """Spiderman is surfing""" UpperCAmelCase = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type="""pt""" ).frames UpperCAmelCase = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
673
1
"""simple docstring""" from __future__ import annotations import os from typing import Any import requests lowerCAmelCase_ : Optional[int] = '''https://api.github.com''' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user lowerCAmelCase_ : Optional[int] = BASE_URL + '''/user''' # https://github.com/settings/tokens lowerCAmelCase_ : int = os.environ.get('''USER_TOKEN''', '''''') def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = { """Authorization""": F'''token {auth_token}''', """Accept""": """application/vnd.github.v3+json""", } return requests.get(lowerCAmelCase , headers=lowerCAmelCase ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F'{key}: {value}') else: raise ValueError('''\'USER_TOKEN\' field cannot be empty.''')
673
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class UpperCamelCase_ ( a_ ): _A : int = 'wav2vec2' def __init__( self , snake_case__=32 , snake_case__=7_68 , snake_case__=12 , snake_case__=12 , snake_case__=30_72 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=1e-5 , snake_case__="group" , snake_case__="gelu" , snake_case__=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case__=(5, 2, 2, 2, 2, 2, 2) , snake_case__=(10, 3, 3, 3, 3, 2, 2) , snake_case__=False , snake_case__=1_28 , snake_case__=16 , snake_case__=False , snake_case__=True , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__=3_20 , snake_case__=2 , snake_case__=0.1 , snake_case__=1_00 , snake_case__=2_56 , snake_case__=2_56 , snake_case__=0.1 , snake_case__="sum" , snake_case__=False , snake_case__=False , snake_case__=2_56 , snake_case__=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case__=(5, 3, 3, 1, 1) , snake_case__=(1, 2, 3, 1, 1) , snake_case__=5_12 , snake_case__=0 , snake_case__=1 , snake_case__=2 , snake_case__=False , snake_case__=3 , snake_case__=2 , snake_case__=3 , snake_case__=None , snake_case__=None , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) UpperCAmelCase = hidden_size UpperCAmelCase = feat_extract_norm UpperCAmelCase = feat_extract_activation UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = conv_bias UpperCAmelCase = num_conv_pos_embeddings UpperCAmelCase = num_conv_pos_embedding_groups UpperCAmelCase = len(self.conv_dim ) UpperCAmelCase = num_hidden_layers UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = feat_proj_dropout UpperCAmelCase = final_dropout UpperCAmelCase = layerdrop UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = vocab_size UpperCAmelCase = do_stable_layer_norm UpperCAmelCase = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase = apply_spec_augment UpperCAmelCase = mask_time_prob UpperCAmelCase = mask_time_length UpperCAmelCase = mask_time_min_masks UpperCAmelCase = mask_feature_prob UpperCAmelCase = mask_feature_length UpperCAmelCase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCAmelCase = num_codevectors_per_group UpperCAmelCase = num_codevector_groups UpperCAmelCase = contrastive_logits_temperature UpperCAmelCase = feat_quantizer_dropout UpperCAmelCase = num_negatives UpperCAmelCase = codevector_dim UpperCAmelCase = proj_codevector_dim UpperCAmelCase = diversity_loss_weight # ctc loss UpperCAmelCase = ctc_loss_reduction UpperCAmelCase = ctc_zero_infinity # adapter UpperCAmelCase = add_adapter UpperCAmelCase = adapter_kernel_size UpperCAmelCase = adapter_stride UpperCAmelCase = num_adapter_layers UpperCAmelCase = output_hidden_size or hidden_size UpperCAmelCase = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = xvector_output_dim @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
673
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ : str = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[int] = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowerCAmelCase_ : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCAmelCase_ : Optional[Any] = NewType('''DataClass''', Any) lowerCAmelCase_ : Any = NewType('''DataClassType''', Any) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = {str(lowerCAmelCase ): choice for choice in choices} return lambda lowerCAmelCase : str_to_choice.get(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( *, lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = None , **lowerCAmelCase , ): '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCAmelCase = {} if aliases is not None: UpperCAmelCase = aliases if help is not None: UpperCAmelCase = help return dataclasses.field(metadata=lowerCAmelCase , default=lowerCAmelCase , default_factory=lowerCAmelCase , **lowerCAmelCase ) class UpperCamelCase_ ( a_ ): _A : Iterable[DataClassType] def __init__( self , snake_case__ , **snake_case__ ) -> List[str]: """simple docstring""" if "formatter_class" not in kwargs: UpperCAmelCase = ArgumentDefaultsHelpFormatter super().__init__(**snake_case__ ) if dataclasses.is_dataclass(snake_case__ ): UpperCAmelCase = [dataclass_types] UpperCAmelCase = list(snake_case__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(snake_case__ ) @staticmethod def UpperCamelCase_ ( snake_case__ , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = f'''--{field.name}''' UpperCAmelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , snake_case__ ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) UpperCAmelCase = kwargs.pop("""aliases""" , [] ) if isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [aliases] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(snake_case__ , """UnionType""" ) and isinstance(snake_case__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(snake_case__ ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" f''' Problem encountered in field \'{field.name}\'.''' ) if type(snake_case__ ) not in field.type.__args__: # filter `str` in Union UpperCAmelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCAmelCase = ( field.type.__args__[0] if isinstance(snake_case__ , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCAmelCase = {} if origin_type is Literal or (isinstance(field.type , snake_case__ ) and issubclass(field.type , snake_case__ )): if origin_type is Literal: UpperCAmelCase = field.type.__args__ else: UpperCAmelCase = [x.value for x in field.type] UpperCAmelCase = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default else: UpperCAmelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCAmelCase = copy(snake_case__ ) # Hack because type=bool in argparse does not behave as we want. UpperCAmelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCAmelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCAmelCase = default # This tells argparse we accept 0 or 1 value after --field_name UpperCAmelCase = """?""" # This is the value that will get picked if we do --field_name (without value) UpperCAmelCase = True elif isclass(snake_case__ ) and issubclass(snake_case__ , snake_case__ ): UpperCAmelCase = field.type.__args__[0] UpperCAmelCase = """+""" if field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() elif field.default is dataclasses.MISSING: UpperCAmelCase = True else: UpperCAmelCase = field.type if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default elif field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() else: UpperCAmelCase = True parser.add_argument(snake_case__ , *snake_case__ , **snake_case__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCAmelCase = False parser.add_argument(f'''--no_{field.name}''' , action="""store_false""" , dest=field.name , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Any: """simple docstring""" if hasattr(snake_case__ , """_argument_group_name""" ): UpperCAmelCase = self.add_argument_group(dtype._argument_group_name ) else: UpperCAmelCase = self try: UpperCAmelCase = get_type_hints(snake_case__ ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(snake_case__ ): UpperCAmelCase = """.""".join(map(snake_case__ , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(snake_case__ ): if not field.init: continue UpperCAmelCase = type_hints[field.name] self._parse_dataclass_field(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self , snake_case__=None , snake_case__=False , snake_case__=True , snake_case__=None , snake_case__=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCAmelCase = [] if args_filename: args_files.append(Path(snake_case__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCAmelCase = ArgumentParser() args_file_parser.add_argument(snake_case__ , type=snake_case__ , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCAmelCase , UpperCAmelCase = args_file_parser.parse_known_args(args=snake_case__ ) UpperCAmelCase = vars(snake_case__ ).get(args_file_flag.lstrip("""-""" ) , snake_case__ ) if cmd_args_file_paths: args_files.extend([Path(snake_case__ ) for p in cmd_args_file_paths] ) UpperCAmelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCAmelCase = file_args + args if args is not None else file_args + sys.argv[1:] UpperCAmelCase , UpperCAmelCase = self.parse_known_args(args=snake_case__ ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in vars(snake_case__ ).items() if k in keys} for k in keys: delattr(snake_case__ , snake_case__ ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(snake_case__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = set(args.keys() ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(snake_case__ )}''' ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(snake_case__ ) , encoding="""utf-8""" ) as open_json_file: UpperCAmelCase = json.loads(open_json_file.read() ) UpperCAmelCase = self.parse_dict(snake_case__ , allow_extra_keys=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = self.parse_dict(yaml.safe_load(Path(snake_case__ ).read_text() ) , allow_extra_keys=snake_case__ ) return tuple(snake_case__ )
673
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : List[Any] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = DPTConfig() if "large" in checkpoint_url: UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 UpperCAmelCase = [5, 11, 17, 23] UpperCAmelCase = [256, 512, 1024, 1024] UpperCAmelCase = (1, 384, 384) if "ade" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = 150 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """ade20k-id2label.json""" UpperCAmelCase = json.load(open(cached_download(hf_hub_url(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = [1, 150, 480, 480] return config, expected_shape def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase = name.replace("""pretrained.model""" , """dpt.encoder""" ) if "pretrained.model" in name: UpperCAmelCase = name.replace("""pretrained.model""" , """dpt.embeddings""" ) if "patch_embed" in name: UpperCAmelCase = name.replace("""patch_embed""" , """patch_embeddings""" ) if "pos_embed" in name: UpperCAmelCase = name.replace("""pos_embed""" , """position_embeddings""" ) if "attn.proj" in name: UpperCAmelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "proj" in name and "project" not in name: UpperCAmelCase = name.replace("""proj""" , """projection""" ) if "blocks" in name: UpperCAmelCase = name.replace("""blocks""" , """layer""" ) if "mlp.fc1" in name: UpperCAmelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: UpperCAmelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "norm1" in name: UpperCAmelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: UpperCAmelCase = name.replace("""norm2""" , """layernorm_after""" ) if "scratch.output_conv" in name: UpperCAmelCase = name.replace("""scratch.output_conv""" , """head""" ) if "scratch" in name: UpperCAmelCase = name.replace("""scratch""" , """neck""" ) if "layer1_rn" in name: UpperCAmelCase = name.replace("""layer1_rn""" , """convs.0""" ) if "layer2_rn" in name: UpperCAmelCase = name.replace("""layer2_rn""" , """convs.1""" ) if "layer3_rn" in name: UpperCAmelCase = name.replace("""layer3_rn""" , """convs.2""" ) if "layer4_rn" in name: UpperCAmelCase = name.replace("""layer4_rn""" , """convs.3""" ) if "refinenet" in name: UpperCAmelCase = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase = name.replace(F'''refinenet{layer_idx}''' , F'''fusion_stage.layers.{abs(layer_idx-4 )}''' ) if "out_conv" in name: UpperCAmelCase = name.replace("""out_conv""" , """projection""" ) if "resConfUnit1" in name: UpperCAmelCase = name.replace("""resConfUnit1""" , """residual_layer1""" ) if "resConfUnit2" in name: UpperCAmelCase = name.replace("""resConfUnit2""" , """residual_layer2""" ) if "conv1" in name: UpperCAmelCase = name.replace("""conv1""" , """convolution1""" ) if "conv2" in name: UpperCAmelCase = name.replace("""conv2""" , """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess1.0.project.0""" , """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess2.0.project.0""" , """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess3.0.project.0""" , """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess4.0.project.0""" , """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess1.3""" , """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess1.4""" , """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess2.3""" , """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess2.4""" , """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess3.3""" , """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess4.3""" , """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase = name.replace("""pretrained.act_postprocess4.4""" , """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: UpperCAmelCase = name.replace("""pretrained""" , """dpt""" ) if "bn" in name: UpperCAmelCase = name.replace("""bn""" , """batch_norm""" ) if "head" in name: UpperCAmelCase = name.replace("""head""" , """head.head""" ) if "encoder.norm" in name: UpperCAmelCase = name.replace("""encoder.norm""" , """layernorm""" ) if "auxlayer" in name: UpperCAmelCase = name.replace("""auxlayer""" , """auxiliary_head.head""" ) return name def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[: config.hidden_size, :] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = get_dpt_config(lowerCAmelCase ) # load original state_dict from URL UpperCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase , map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase = state_dict.pop(lowerCAmelCase ) UpperCAmelCase = val # read in qkv matrices read_in_q_k_v(lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model UpperCAmelCase = DPTForSemanticSegmentation(lowerCAmelCase ) if """ade""" in checkpoint_url else DPTForDepthEstimation(lowerCAmelCase ) model.load_state_dict(lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase = 480 if """ade""" in checkpoint_url else 384 UpperCAmelCase = DPTImageProcessor(size=lowerCAmelCase ) UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(lowerCAmelCase , return_tensors="""pt""" ) # forward pass UpperCAmelCase = model(**lowerCAmelCase ).logits if """ade""" in checkpoint_url else model(**lowerCAmelCase ).predicted_depth # Assert logits UpperCAmelCase = torch.tensor([[6.31_99, 6.36_29, 6.41_48], [6.38_50, 6.36_15, 6.41_66], [6.35_19, 6.31_76, 6.35_75]] ) if "ade" in checkpoint_url: UpperCAmelCase = torch.tensor([[4.04_80, 4.24_20, 4.43_60], [4.31_24, 4.56_93, 4.82_61], [4.57_68, 4.89_65, 5.21_63]] ) assert outputs.shape == torch.Size(lowerCAmelCase ) assert ( torch.allclose(outputs[0, 0, :3, :3] , lowerCAmelCase , atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , lowerCAmelCase ) ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(lowerCAmelCase , lowerCAmelCase ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=lowerCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(lowerCAmelCase , lowerCAmelCase ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=lowerCAmelCase , ) if __name__ == "__main__": lowerCAmelCase_ : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) lowerCAmelCase_ : Optional[Any] = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
673
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCAmelCase_ : List[str] = False class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self , snake_case__=32 ) -> Optional[Any]: """simple docstring""" set_seed(0 ) UpperCAmelCase = UNetaDModel(sample_size=snake_case__ , in_channels=3 , out_channels=3 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) UpperCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randint(0 , 10_00 , (4,) ).long().to(snake_case__ ) for _ in range(4 )] # train with a DDPM scheduler UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) )
673
1
"""simple docstring""" # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() lowerCAmelCase_ : Union[str, Any] = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model lowerCAmelCase_ : str = { # fairseq: '''wmt19-ru-en''': {'''length_penalty''': 1.1}, '''wmt19-en-ru''': {'''length_penalty''': 1.15}, '''wmt19-en-de''': {'''length_penalty''': 1.0}, '''wmt19-de-en''': {'''length_penalty''': 1.1}, # allenai: '''wmt16-en-de-dist-12-1''': {'''length_penalty''': 0.6}, '''wmt16-en-de-dist-6-1''': {'''length_penalty''': 0.6}, '''wmt16-en-de-12-1''': {'''length_penalty''': 0.8}, '''wmt19-de-en-6-6-base''': {'''length_penalty''': 0.6}, '''wmt19-de-en-6-6-big''': {'''length_penalty''': 0.6}, } # this remaps the different models to their organization names lowerCAmelCase_ : List[str] = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowerCAmelCase_ : str = '''facebook''' for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: lowerCAmelCase_ : Any = '''allenai''' def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} UpperCAmelCase = dict((re.sub(r"""@@$""" , """""" , lowerCAmelCase ), v) if k.endswith("""@@""" ) else (re.sub(r"""$""" , """</w>""" , lowerCAmelCase ), v) for k, v in d.items() ) UpperCAmelCase = """<s> <pad> </s> <unk>""".split() # restore the special tokens for k in keep_keys: del da[F'''{k}</w>'''] UpperCAmelCase = d[k] # restore return da def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' # prep assert os.path.exists(lowerCAmelCase ) os.makedirs(lowerCAmelCase , exist_ok=lowerCAmelCase ) print(F'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models UpperCAmelCase = basename(lowerCAmelCase ) UpperCAmelCase = dirname(lowerCAmelCase ) UpperCAmelCase = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel UpperCAmelCase = cls.hub_models() UpperCAmelCase = {"""bpe""": """fastbpe""", """tokenizer""": """moses"""} UpperCAmelCase = """.""" # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(F'''using checkpoint {checkpoint_file}''' ) UpperCAmelCase = hub_utils.from_pretrained( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , archive_map=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase = vars(chkpt["""args"""]["""model"""] ) UpperCAmelCase = args["""source_lang"""] UpperCAmelCase = args["""target_lang"""] UpperCAmelCase = dirname(lowerCAmelCase ) UpperCAmelCase = basename(lowerCAmelCase ) # dicts UpperCAmelCase = os.path.join(lowerCAmelCase , F'''dict.{src_lang}.txt''' ) UpperCAmelCase = os.path.join(lowerCAmelCase , F'''dict.{tgt_lang}.txt''' ) UpperCAmelCase = Dictionary.load(lowerCAmelCase ) UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) UpperCAmelCase = len(lowerCAmelCase ) UpperCAmelCase = os.path.join(lowerCAmelCase , """vocab-src.json""" ) print(F'''Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records''' ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(lowerCAmelCase , ensure_ascii=lowerCAmelCase , indent=lowerCAmelCase ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab UpperCAmelCase = True for k in src_vocab.keys(): if not k.islower(): UpperCAmelCase = False break UpperCAmelCase = Dictionary.load(lowerCAmelCase ) UpperCAmelCase = rewrite_dict_keys(tgt_dict.indices ) UpperCAmelCase = len(lowerCAmelCase ) UpperCAmelCase = os.path.join(lowerCAmelCase , """vocab-tgt.json""" ) print(F'''Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records''' ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(lowerCAmelCase , ensure_ascii=lowerCAmelCase , indent=lowerCAmelCase ) ) # merges_file (bpecodes) UpperCAmelCase = os.path.join(lowerCAmelCase , VOCAB_FILES_NAMES["""merges_file"""] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" UpperCAmelCase = os.path.join(lowerCAmelCase , lowerCAmelCase ) if os.path.exists(lowerCAmelCase ): break with open(lowerCAmelCase , encoding="""utf-8""" ) as fin: UpperCAmelCase = fin.read() UpperCAmelCase = re.sub(r""" \d+$""" , """""" , lowerCAmelCase , 0 , re.M ) # remove frequency number print(F'''Generating {merges_file}''' ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as fout: fout.write(lowerCAmelCase ) # model config UpperCAmelCase = os.path.join(lowerCAmelCase , """config.json""" ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", F'''need to extend tokenizer to support bpe={args['bpe']}''' assert args["tokenizer"] == "moses", F'''need to extend tokenizer to support bpe={args['tokenizer']}''' UpperCAmelCase = { """architectures""": ["""FSMTForConditionalGeneration"""], """model_type""": """fsmt""", """activation_dropout""": args["""activation_dropout"""], """activation_function""": """relu""", """attention_dropout""": args["""attention_dropout"""], """d_model""": args["""decoder_embed_dim"""], """dropout""": args["""dropout"""], """init_std""": 0.02, """max_position_embeddings""": args["""max_source_positions"""], """num_hidden_layers""": args["""encoder_layers"""], """src_vocab_size""": src_vocab_size, """tgt_vocab_size""": tgt_vocab_size, """langs""": [src_lang, tgt_lang], """encoder_attention_heads""": args["""encoder_attention_heads"""], """encoder_ffn_dim""": args["""encoder_ffn_embed_dim"""], """encoder_layerdrop""": args["""encoder_layerdrop"""], """encoder_layers""": args["""encoder_layers"""], """decoder_attention_heads""": args["""decoder_attention_heads"""], """decoder_ffn_dim""": args["""decoder_ffn_embed_dim"""], """decoder_layerdrop""": args["""decoder_layerdrop"""], """decoder_layers""": args["""decoder_layers"""], """bos_token_id""": 0, """pad_token_id""": 1, """eos_token_id""": 2, """is_encoder_decoder""": True, """scale_embedding""": not args["""no_scale_embedding"""], """tie_word_embeddings""": args["""share_all_embeddings"""], } # good hparam defaults to start with UpperCAmelCase = 5 UpperCAmelCase = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: UpperCAmelCase = best_score_hparams[model_dir]["""length_penalty"""] else: UpperCAmelCase = 1.0 print(F'''Generating {fsmt_model_config_file}''' ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(lowerCAmelCase , ensure_ascii=lowerCAmelCase , indent=lowerCAmelCase ) ) # tokenizer config UpperCAmelCase = os.path.join(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase = { """langs""": [src_lang, tgt_lang], """model_max_length""": 1024, """do_lower_case""": do_lower_case, } print(F'''Generating {fsmt_tokenizer_config_file}''' ) with open(lowerCAmelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(lowerCAmelCase , ensure_ascii=lowerCAmelCase , indent=lowerCAmelCase ) ) # model UpperCAmelCase = chkpt["""models"""][0] UpperCAmelCase = model.state_dict() # rename keys to start with 'model.' UpperCAmelCase = OrderedDict(("""model.""" + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys UpperCAmelCase = [ """model.model""", """model.encoder.version""", """model.decoder.version""", """model.encoder_embed_tokens.weight""", """model.decoder_embed_tokens.weight""", """model.encoder.embed_positions._float_tensor""", """model.decoder.embed_positions._float_tensor""", ] for k in ignore_keys: model_state_dict.pop(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase = FSMTConfig.from_pretrained(lowerCAmelCase ) UpperCAmelCase = FSMTForConditionalGeneration(lowerCAmelCase ) # check that it loads ok model_new.load_state_dict(lowerCAmelCase , strict=lowerCAmelCase ) # save UpperCAmelCase = os.path.join(lowerCAmelCase , lowerCAmelCase ) print(F'''Generating {pytorch_weights_dump_path}''' ) torch.save(lowerCAmelCase , lowerCAmelCase ) print("""Conversion is done!""" ) print("""\nLast step is to upload the files to s3""" ) print(F'''cd {data_root}''' ) print(F'''transformers-cli upload {model_dir}''' ) if __name__ == "__main__": lowerCAmelCase_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fsmt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase_ : int = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
673
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class UpperCamelCase_ : def __init__( self , snake_case__=2 , snake_case__=3 , snake_case__=64 , snake_case__=None ) -> List[str]: """simple docstring""" UpperCAmelCase = np.random.default_rng(snake_case__ ) UpperCAmelCase = length UpperCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) UpperCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> int: """simple docstring""" return self.length def __getitem__( self , snake_case__ ) -> Tuple: """simple docstring""" return {"x": self.x[i], "y": self.y[i]} class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[str]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> List[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a[0] + self.b[0] class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> Optional[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a + self.b def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = 16 ): '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer UpperCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) UpperCAmelCase = {"""train""": """tests/test_samples/MRPC/train.csv""", """validation""": """tests/test_samples/MRPC/dev.csv"""} UpperCAmelCase = load_dataset("""csv""" , data_files=lowerCAmelCase ) UpperCAmelCase = datasets["""train"""].unique("""label""" ) UpperCAmelCase = {v: i for i, v in enumerate(lowerCAmelCase )} def tokenize_function(lowerCAmelCase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase = tokenizer( examples["""sentence1"""] , examples["""sentence2"""] , truncation=lowerCAmelCase , max_length=lowerCAmelCase , padding="""max_length""" ) if "label" in examples: UpperCAmelCase = [label_to_id[l] for l in examples["""label"""]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase = datasets.map( lowerCAmelCase , batched=lowerCAmelCase , remove_columns=["""sentence1""", """sentence2""", """label"""] , ) def collate_fn(lowerCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return tokenizer.pad(lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. UpperCAmelCase = DataLoader(tokenized_datasets["""train"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=2 ) UpperCAmelCase = DataLoader(tokenized_datasets["""validation"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
673
1
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = len(lowerCAmelCase ) + 1 UpperCAmelCase = len(lowerCAmelCase ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. UpperCAmelCase = [[0 for i in range(lowerCAmelCase )] for j in range(lowerCAmelCase )] # since string of zero length match pattern of zero length UpperCAmelCase = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , lowerCAmelCase ): UpperCAmelCase = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , lowerCAmelCase ): UpperCAmelCase = dp[0][j - 2] if pattern[j - 1] == """*""" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , lowerCAmelCase ): for j in range(1 , lowerCAmelCase ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": UpperCAmelCase = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: UpperCAmelCase = 1 elif pattern[j - 2] in (input_string[i - 1], "."): UpperCAmelCase = dp[i - 1][j] else: UpperCAmelCase = 0 else: UpperCAmelCase = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") lowerCAmelCase_ : str = '''aab''' lowerCAmelCase_ : Optional[Any] = '''c*a*b''' # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F'{input_string} matches the given pattern {pattern}') else: print(F'{input_string} does not match with the given pattern {pattern}')
673
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = hidden_states.shape UpperCAmelCase = jax.image.resize( snake_case__ , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Any: """simple docstring""" UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : int = None _A : float = 0.0 _A : bool = None _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.in_channels if self.out_channels is None else self.out_channels UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = nn.Dense(snake_case__ , dtype=self.dtype ) UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Dropout(self.dropout_prob ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut UpperCAmelCase = None if use_nin_shortcut: UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self , snake_case__ , snake_case__ , snake_case__=True ) -> List[Any]: """simple docstring""" UpperCAmelCase = hidden_states UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) UpperCAmelCase = self.time_emb_proj(nn.swish(snake_case__ ) ) UpperCAmelCase = jnp.expand_dims(jnp.expand_dims(snake_case__ , 1 ) , 1 ) UpperCAmelCase = hidden_states + temb UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.dropout(snake_case__ , snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) if self.conv_shortcut is not None: UpperCAmelCase = self.conv_shortcut(snake_case__ ) return hidden_states + residual
673
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: lowerCAmelCase_ : str = None lowerCAmelCase_ : List[str] = logging.get_logger(__name__) lowerCAmelCase_ : Optional[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase_ : Any = { '''vocab_file''': { '''google/bigbird-roberta-base''': '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model''', '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model''' ), }, '''tokenizer_file''': { '''google/bigbird-roberta-base''': ( '''https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json''' ), '''google/bigbird-roberta-large''': ( '''https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json''' ), '''google/bigbird-base-trivia-itc''': ( '''https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase_ : Tuple = { '''google/bigbird-roberta-base''': 4_0_9_6, '''google/bigbird-roberta-large''': 4_0_9_6, '''google/bigbird-base-trivia-itc''': 4_0_9_6, } lowerCAmelCase_ : int = '''▁''' class UpperCamelCase_ ( a_ ): _A : List[str] = VOCAB_FILES_NAMES _A : Any = PRETRAINED_VOCAB_FILES_MAP _A : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : Optional[Any] = BigBirdTokenizer _A : Any = ['input_ids', 'attention_mask'] _A : List[int] = [] def __init__( self , snake_case__=None , snake_case__=None , snake_case__="<unk>" , snake_case__="<s>" , snake_case__="</s>" , snake_case__="<pad>" , snake_case__="[SEP]" , snake_case__="[MASK]" , snake_case__="[CLS]" , **snake_case__ , ) -> Any: """simple docstring""" UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else bos_token UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else eos_token UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else unk_token UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else pad_token UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else cls_token UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( snake_case__ , tokenizer_file=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , **snake_case__ , ) UpperCAmelCase = vocab_file UpperCAmelCase = False if not self.vocab_file else True def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None , snake_case__ = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( """You should not supply a second sequence if the provided sequence of """ """ids is already formatted with special tokens for the model.""" ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1] + ([0] * len(snake_case__ )) + [1] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> List[int]: """simple docstring""" UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self , snake_case__ , snake_case__ = None ) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(snake_case__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase = os.path.join( snake_case__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
673
"""simple docstring""" from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=30 , snake_case__=2 , snake_case__=3 , snake_case__=True , snake_case__=True , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=10 , snake_case__=0.02 , snake_case__=3 , snake_case__=None , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 1 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModel(config=snake_case__ ) UpperCAmelCase = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) UpperCAmelCase = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[int] = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _A : Optional[Any] = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) _A : Optional[int] = False _A : Any = False _A : List[str] = False def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(snake_case__ ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=snake_case__ , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**snake_case__ ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case__ ) UpperCAmelCase = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , snake_case__ , atol=1e-4 )
673
1
def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" return round(float(moles / volume ) * nfactor ) def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" return round(float((moles * 0.0821 * temperature) / (volume) ) ) def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
0
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=5_12 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = NystromformerForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = NystromformerForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = self.num_choices UpperCAmelCase = NystromformerForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[Any] = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) _A : Optional[Any] = ( { 'feature-extraction': NystromformerModel, 'fill-mask': NystromformerForMaskedLM, 'question-answering': NystromformerForQuestionAnswering, 'text-classification': NystromformerForSequenceClassification, 'token-classification': NystromformerForTokenClassification, 'zero-shot': NystromformerForSequenceClassification, } if is_torch_available() else {} ) _A : int = False _A : Dict = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = NystromformerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case__ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = NystromformerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): UpperCAmelCase = model(snake_case__ )[0] UpperCAmelCase = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """the [MASK] of Belgium is Brussels""" UpperCAmelCase = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = tokenizer(snake_case__ , return_tensors="""pt""" ) with torch.no_grad(): UpperCAmelCase = model(encoding.input_ids ).logits UpperCAmelCase = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(snake_case__ ) , """capital""" )
673
0
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class __lowerCamelCase (_a ): _lowercase = ["""image_processor"""] _lowercase = """SamImageProcessor""" def __init__( self: Dict,A_: Union[str, Any] ): '''simple docstring''' super().__init__(A_ ) __UpperCamelCase = self.image_processor __UpperCamelCase = -10 __UpperCamelCase = self.image_processor.size['longest_edge'] def __call__( self: Optional[Any],A_: Optional[int]=None,A_: int=None,A_: str=None,A_: str=None,A_: Optional[Union[str, TensorType]] = None,**A_: Optional[int],): '''simple docstring''' __UpperCamelCase = self.image_processor( A_,return_tensors=A_,**A_,) # pop arguments that are not used in the foward but used nevertheless __UpperCamelCase = encoding_image_processor['original_sizes'] if hasattr(A_,'numpy' ): # Checks if Torch or TF tensor __UpperCamelCase = original_sizes.numpy() __UpperCamelCase, __UpperCamelCase, __UpperCamelCase = self._check_and_preprocess_points( input_points=A_,input_labels=A_,input_boxes=A_,) __UpperCamelCase = self._normalize_and_convert( A_,A_,input_points=A_,input_labels=A_,input_boxes=A_,return_tensors=A_,) return encoding_image_processor def snake_case_ ( self: Tuple,A_: Any,A_: str,A_: Dict=None,A_: Dict=None,A_: int=None,A_: List[Any]="pt",): '''simple docstring''' if input_points is not None: if len(A_ ) != len(A_ ): __UpperCamelCase = [ self._normalize_coordinates(self.target_size,A_,original_sizes[0] ) for point in input_points ] else: __UpperCamelCase = [ self._normalize_coordinates(self.target_size,A_,A_ ) for point, original_size in zip(A_,A_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: __UpperCamelCase, __UpperCamelCase = self._pad_points_and_labels(A_,A_ ) __UpperCamelCase = np.array(A_ ) if input_labels is not None: __UpperCamelCase = np.array(A_ ) if input_boxes is not None: if len(A_ ) != len(A_ ): __UpperCamelCase = [ self._normalize_coordinates(self.target_size,A_,original_sizes[0],is_bounding_box=A_ ) for box in input_boxes ] else: __UpperCamelCase = [ self._normalize_coordinates(self.target_size,A_,A_,is_bounding_box=A_ ) for box, original_size in zip(A_,A_ ) ] __UpperCamelCase = np.array(A_ ) if input_boxes is not None: if return_tensors == "pt": __UpperCamelCase = torch.from_numpy(A_ ) # boxes batch size of 1 by default __UpperCamelCase = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": __UpperCamelCase = tf.convert_to_tensor(A_ ) # boxes batch size of 1 by default __UpperCamelCase = tf.expand_dims(A_,1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'input_boxes': input_boxes} ) if input_points is not None: if return_tensors == "pt": __UpperCamelCase = torch.from_numpy(A_ ) # point batch size of 1 by default __UpperCamelCase = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": __UpperCamelCase = tf.convert_to_tensor(A_ ) # point batch size of 1 by default __UpperCamelCase = tf.expand_dims(A_,1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'input_points': input_points} ) if input_labels is not None: if return_tensors == "pt": __UpperCamelCase = torch.from_numpy(A_ ) # point batch size of 1 by default __UpperCamelCase = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": __UpperCamelCase = tf.convert_to_tensor(A_ ) # point batch size of 1 by default __UpperCamelCase = tf.expand_dims(A_,1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'input_labels': input_labels} ) return encoding_image_processor def snake_case_ ( self: List[str],A_: int,A_: Optional[Any] ): '''simple docstring''' __UpperCamelCase = max([point.shape[0] for point in input_points] ) __UpperCamelCase = [] for i, point in enumerate(A_ ): if point.shape[0] != expected_nb_points: __UpperCamelCase = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value],axis=0 ) __UpperCamelCase = np.append(input_labels[i],[self.point_pad_value] ) processed_input_points.append(A_ ) __UpperCamelCase = processed_input_points return input_points, input_labels def snake_case_ ( self: Optional[int],A_: int,A_: np.ndarray,A_: Union[str, Any],A_: int=False ): '''simple docstring''' __UpperCamelCase, __UpperCamelCase = original_size __UpperCamelCase, __UpperCamelCase = self.image_processor._get_preprocess_shape(A_,longest_edge=A_ ) __UpperCamelCase = deepcopy(A_ ).astype(A_ ) if is_bounding_box: __UpperCamelCase = coords.reshape(-1,2,2 ) __UpperCamelCase = coords[..., 0] * (new_w / old_w) __UpperCamelCase = coords[..., 1] * (new_h / old_h) if is_bounding_box: __UpperCamelCase = coords.reshape(-1,4 ) return coords def snake_case_ ( self: Dict,A_: Optional[Any]=None,A_: List[str]=None,A_: Tuple=None,): '''simple docstring''' if input_points is not None: if hasattr(A_,'numpy' ): # Checks for TF or Torch tensor __UpperCamelCase = input_points.numpy().tolist() if not isinstance(A_,A_ ) or not isinstance(input_points[0],A_ ): raise ValueError('Input points must be a list of list of floating points.' ) __UpperCamelCase = [np.array(A_ ) for input_point in input_points] else: __UpperCamelCase = None if input_labels is not None: if hasattr(A_,'numpy' ): __UpperCamelCase = input_labels.numpy().tolist() if not isinstance(A_,A_ ) or not isinstance(input_labels[0],A_ ): raise ValueError('Input labels must be a list of list integers.' ) __UpperCamelCase = [np.array(A_ ) for label in input_labels] else: __UpperCamelCase = None if input_boxes is not None: if hasattr(A_,'numpy' ): __UpperCamelCase = input_boxes.numpy().tolist() if ( not isinstance(A_,A_ ) or not isinstance(input_boxes[0],A_ ) or not isinstance(input_boxes[0][0],A_ ) ): raise ValueError('Input boxes must be a list of list of list of floating points.' ) __UpperCamelCase = [np.array(A_ ).astype(np.floataa ) for box in input_boxes] else: __UpperCamelCase = None return input_points, input_labels, input_boxes @property def snake_case_ ( self: List[str] ): '''simple docstring''' __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(A_ ) ) def snake_case_ ( self: int,*A_: int,**A_: Tuple ): '''simple docstring''' return self.image_processor.post_process_masks(*A_,**A_ )
1
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Optional[int] = False def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return TrainCommand(lowerCAmelCase ) class UpperCamelCase_ ( a_ ): @staticmethod def UpperCamelCase_ ( snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = parser.add_parser("""train""" , help="""CLI tool to train a model on a task.""" ) train_parser.add_argument( """--train_data""" , type=snake_case__ , required=snake_case__ , help="""path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.""" , ) train_parser.add_argument( """--column_label""" , type=snake_case__ , default=0 , help="""Column of the dataset csv file with example labels.""" ) train_parser.add_argument( """--column_text""" , type=snake_case__ , default=1 , help="""Column of the dataset csv file with example texts.""" ) train_parser.add_argument( """--column_id""" , type=snake_case__ , default=2 , help="""Column of the dataset csv file with example ids.""" ) train_parser.add_argument( """--skip_first_row""" , action="""store_true""" , help="""Skip the first row of the csv file (headers).""" ) train_parser.add_argument("""--validation_data""" , type=snake_case__ , default="""""" , help="""path to validation dataset.""" ) train_parser.add_argument( """--validation_split""" , type=snake_case__ , default=0.1 , help="""if validation dataset is not provided, fraction of train dataset to use as validation dataset.""" , ) train_parser.add_argument("""--output""" , type=snake_case__ , default="""./""" , help="""path to saved the trained model.""" ) train_parser.add_argument( """--task""" , type=snake_case__ , default="""text_classification""" , help="""Task to train the model on.""" ) train_parser.add_argument( """--model""" , type=snake_case__ , default="""bert-base-uncased""" , help="""Model's name or path to stored model.""" ) train_parser.add_argument("""--train_batch_size""" , type=snake_case__ , default=32 , help="""Batch size for training.""" ) train_parser.add_argument("""--valid_batch_size""" , type=snake_case__ , default=64 , help="""Batch size for validation.""" ) train_parser.add_argument("""--learning_rate""" , type=snake_case__ , default=3e-5 , help="""Learning rate.""" ) train_parser.add_argument("""--adam_epsilon""" , type=snake_case__ , default=1e-08 , help="""Epsilon for Adam optimizer.""" ) train_parser.set_defaults(func=snake_case__ ) def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = logging.get_logger("""transformers-cli/training""" ) UpperCAmelCase = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=snake_case__ ) UpperCAmelCase = args.output UpperCAmelCase = args.column_label UpperCAmelCase = args.column_text UpperCAmelCase = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": UpperCAmelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = args.validation_split UpperCAmelCase = args.train_batch_size UpperCAmelCase = args.valid_batch_size UpperCAmelCase = args.learning_rate UpperCAmelCase = args.adam_epsilon def UpperCamelCase_ ( self ) -> Any: """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" raise NotImplementedError def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
673
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCamelCase__ ( _A): """simple docstring""" a__ : Any = "ctrl" a__ : Tuple = ["past_key_values"] a__ : Union[str, Any] = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : List[Any] , __lowerCAmelCase : Tuple=24_65_34 , __lowerCAmelCase : str=2_56 , __lowerCAmelCase : int=12_80 , __lowerCAmelCase : int=81_92 , __lowerCAmelCase : str=48 , __lowerCAmelCase : List[Any]=16 , __lowerCAmelCase : int=0.1 , __lowerCAmelCase : Tuple=0.1 , __lowerCAmelCase : List[Any]=1E-6 , __lowerCAmelCase : List[Any]=0.02 , __lowerCAmelCase : str=True , **__lowerCAmelCase : List[Any] , ) -> List[str]: _A = vocab_size _A = n_positions _A = n_embd _A = n_layer _A = n_head _A = dff _A = resid_pdrop _A = embd_pdrop _A = layer_norm_epsilon _A = initializer_range _A = use_cache super().__init__(**__lowerCAmelCase )
2
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=sys.maxsize ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """bilinear""" UpperCAmelCase = max_size UpperCAmelCase = short_edge_length def __call__( self , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = [] for img in imgs: UpperCAmelCase , UpperCAmelCase = img.shape[:2] # later: provide list and randomly choose index for resize UpperCAmelCase = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img UpperCAmelCase = size * 1.0 / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size if max(snake_case__ , snake_case__ ) > self.max_size: UpperCAmelCase = self.max_size * 1.0 / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase = int(neww + 0.5 ) UpperCAmelCase = int(newh + 0.5 ) if img.dtype == np.uinta: UpperCAmelCase = Image.fromarray(snake_case__ ) UpperCAmelCase = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) UpperCAmelCase = np.asarray(snake_case__ ) else: UpperCAmelCase = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw UpperCAmelCase = nn.functional.interpolate( snake_case__ , (newh, neww) , mode=self.interp_method , align_corners=snake_case__ ).squeeze(0 ) img_augs.append(snake_case__ ) return img_augs class UpperCamelCase_ : def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) UpperCAmelCase = cfg.INPUT.FORMAT UpperCAmelCase = cfg.SIZE_DIVISIBILITY UpperCAmelCase = cfg.PAD_VALUE UpperCAmelCase = cfg.INPUT.MAX_SIZE_TEST UpperCAmelCase = cfg.MODEL.DEVICE UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = lambda snake_case__ : (x - self.pixel_mean) / self.pixel_std def UpperCamelCase_ ( self , snake_case__ ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = tuple(max(snake_case__ ) for s in zip(*[img.shape for img in images] ) ) UpperCAmelCase = [im.shape[-2:] for im in images] UpperCAmelCase = [ nn.functional.pad( snake_case__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(snake_case__ , snake_case__ ) ] return torch.stack(snake_case__ ), torch.tensor(snake_case__ ) def __call__( self , snake_case__ , snake_case__=False ) -> Optional[Any]: """simple docstring""" with torch.no_grad(): if not isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [images] if single_image: assert len(snake_case__ ) == 1 for i in range(len(snake_case__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(snake_case__ , images.pop(snake_case__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( snake_case__ , torch.as_tensor(img_tensorize(images.pop(snake_case__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge UpperCAmelCase = torch.tensor([im.shape[:2] for im in images] ) UpperCAmelCase = self.aug(snake_case__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic UpperCAmelCase = [self.normalizer(snake_case__ ) for x in images] # now pad them to do the following operations UpperCAmelCase , UpperCAmelCase = self.pad(snake_case__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad UpperCAmelCase = torch.true_divide(snake_case__ , snake_case__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' assert torch.isfinite(lowerCAmelCase ).all(), "Box tensor contains infinite or NaN!" UpperCAmelCase , UpperCAmelCase = box_size tensor[:, 0].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 1].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 2].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 3].clamp_(min=0 , max=lowerCAmelCase )
673
0
'''simple docstring''' from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def A_( A : Any , A : str): UpperCamelCase = [] for part_id in partition_order: UpperCamelCase = df.where(f'''SPARK_PARTITION_ID() = {part_id}''').collect() for row_idx, row in enumerate(A): expected_row_ids_and_row_dicts.append((f'''{part_id}_{row_idx}''', row.asDict())) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(100).repartition(1) UpperCamelCase = Spark(A) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(10).repartition(2) UpperCamelCase = [1, 0] UpperCamelCase = _generate_iterable_examples(A , A) # Reverse the partitions. UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(A , A) for i, (row_id, row_dict) in enumerate(generate_fn()): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(10).repartition(1) UpperCamelCase = SparkExamplesIterable(A) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(A): assert row_id == f'''0_{i}''' assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(30).repartition(3) # Mock the generator so that shuffle reverses the partition indices. with patch('numpy.random.Generator') as generator_mock: UpperCamelCase = lambda A: x.reverse() UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(A , [2, 1, 0]) UpperCamelCase = SparkExamplesIterable(A).shuffle_data_sources(A) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(A): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(20).repartition(4) # Partitions 0 and 2 UpperCamelCase = SparkExamplesIterable(A).shard_data_sources(worker_id=0 , num_workers=2) assert shard_it_a.n_shards == 2 UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(A , [0, 2]) for i, (row_id, row_dict) in enumerate(A): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 UpperCamelCase = SparkExamplesIterable(A).shard_data_sources(worker_id=1 , num_workers=2) assert shard_it_a.n_shards == 2 UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(A , [1, 3]) for i, (row_id, row_dict) in enumerate(A): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def A_( ): UpperCamelCase = pyspark.sql.SparkSession.builder.master('local[*]').appName('pyspark').getOrCreate() UpperCamelCase = spark.range(100).repartition(1) UpperCamelCase = Spark(A) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
3
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : List[str] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase = """""" else: UpperCAmelCase = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = dct.pop(lowerCAmelCase ) UpperCAmelCase = val def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase = 1000 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = int(deit_name[-6:-4] ) UpperCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): UpperCAmelCase = 192 UpperCAmelCase = 768 UpperCAmelCase = 12 UpperCAmelCase = 3 elif deit_name[9:].startswith("""small""" ): UpperCAmelCase = 384 UpperCAmelCase = 1536 UpperCAmelCase = 12 UpperCAmelCase = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 # load original model from timm UpperCAmelCase = timm.create_model(lowerCAmelCase , pretrained=lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase = timm_model.state_dict() UpperCAmelCase = create_rename_keys(lowerCAmelCase , lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) read_in_q_k_v(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model UpperCAmelCase = DeiTForImageClassificationWithTeacher(lowerCAmelCase ).eval() model.load_state_dict(lowerCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase = DeiTImageProcessor(size=lowerCAmelCase , crop_size=config.image_size ) UpperCAmelCase = image_processor(images=prepare_img() , return_tensors="""pt""" ) UpperCAmelCase = encoding["""pixel_values"""] UpperCAmelCase = model(lowerCAmelCase ) UpperCAmelCase = timm_model(lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase_ : str = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
673
0
"""simple docstring""" # A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str ): lowerCAmelCase = [False] * len(_UpperCAmelCase ) lowerCAmelCase = [-1] * len(_UpperCAmelCase ) def dfs(_UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ): lowerCAmelCase = True lowerCAmelCase = c for u in graph[v]: if not visited[u]: dfs(_UpperCAmelCase , 1 - c ) for i in range(len(_UpperCAmelCase ) ): if not visited[i]: dfs(_UpperCAmelCase , 0 ) for i in range(len(_UpperCAmelCase ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph __UpperCamelCase : Dict = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
4
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = 32 , snake_case__ = True , snake_case__ = 1 / 2_55 , snake_case__ = True , snake_case__ = True , snake_case__ = [0.48_145_466, 0.4_578_275, 0.40_821_073] , snake_case__ = [0.26_862_954, 0.26_130_258, 0.27_577_711] , snake_case__ = True , snake_case__=7 , snake_case__=30 , snake_case__=4_00 , snake_case__=3 , ) -> List[str]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = do_resize UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_88} UpperCAmelCase = size_divisor UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_pad UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase_ ( self , snake_case__ , snake_case__=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = image_inputs[0] if isinstance(snake_case__ , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] UpperCAmelCase = size / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size UpperCAmelCase = int((13_33 / 8_00) * size ) if max(snake_case__ , snake_case__ ) > max_size: UpperCAmelCase = max_size / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase , UpperCAmelCase = int(newh + 0.5 ), int(neww + 0.5 ) UpperCAmelCase , UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """size_divisor""" ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
673
0
'''simple docstring''' from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline _lowercase = logging.get_logger(__name__) @add_end_docstrings(_SCREAMING_SNAKE_CASE ) class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , **_lowercase ): """simple docstring""" super().__init__(**_lowercase ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) # No specific FOR_XXX available yet def __call__( self , _lowercase , **_lowercase ): """simple docstring""" return super().__call__(_lowercase , **_lowercase ) def _lowercase ( self , **_lowercase ): """simple docstring""" _lowerCAmelCase = {} if "candidate_labels" in kwargs: _lowerCAmelCase = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: _lowerCAmelCase = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def _lowercase ( self , _lowercase , _lowercase=None , _lowercase="This is a sound of {}." ): """simple docstring""" if isinstance(_lowercase , _lowercase ): if audio.startswith("""http://""" ) or audio.startswith("""https://""" ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png _lowerCAmelCase = requests.get(_lowercase ).content else: with open(_lowercase , """rb""" ) as f: _lowerCAmelCase = f.read() if isinstance(_lowercase , _lowercase ): _lowerCAmelCase = ffmpeg_read(_lowercase , self.feature_extractor.sampling_rate ) if not isinstance(_lowercase , np.ndarray ): raise ValueError("""We expect a numpy ndarray as input""" ) if len(audio.shape ) != 1: raise ValueError("""We expect a single channel audio input for ZeroShotAudioClassificationPipeline""" ) _lowerCAmelCase = self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors="""pt""" ) _lowerCAmelCase = candidate_labels _lowerCAmelCase = [hypothesis_template.format(_lowercase ) for x in candidate_labels] _lowerCAmelCase = self.tokenizer(_lowercase , return_tensors=self.framework , padding=_lowercase ) _lowerCAmelCase = [text_inputs] return inputs def _lowercase ( self , _lowercase ): """simple docstring""" _lowerCAmelCase = model_inputs.pop("""candidate_labels""" ) _lowerCAmelCase = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] , _lowercase ): _lowerCAmelCase = text_inputs[0] else: # Batching case. _lowerCAmelCase = text_inputs[0][0] _lowerCAmelCase = self.model(**_lowercase , **_lowercase ) _lowerCAmelCase = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_audio, } return model_outputs def _lowercase ( self , _lowercase ): """simple docstring""" _lowerCAmelCase = model_outputs.pop("""candidate_labels""" ) _lowerCAmelCase = model_outputs["""logits"""][0] if self.framework == "pt": _lowerCAmelCase = logits.softmax(dim=0 ) _lowerCAmelCase = probs.tolist() else: raise ValueError("""`tf` framework not supported.""" ) _lowerCAmelCase = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(_lowercase , _lowercase ) , key=lambda _lowercase : -x[0] ) ] return result
5
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[str] = XLMRobertaTokenizer _A : List[str] = XLMRobertaTokenizerFast _A : Optional[Any] = True _A : List[str] = True def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(snake_case__ ) , 10_02 ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(snake_case__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) UpperCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=snake_case__ ) UpperCAmelCase = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = """I was born in 92000, and this is falsé.""" UpperCAmelCase = tokenizer.tokenize(snake_case__ ) UpperCAmelCase = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
673
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() # fmt: off SCREAMING_SNAKE_CASE__ = ["""l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = {"""unk_token""": """<unk>"""} SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__A ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__A ) ) SCREAMING_SNAKE_CASE__ = { """do_resize""": True, """size""": 20, """do_center_crop""": True, """crop_size""": 18, """do_normalize""": True, """image_mean""": [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], """image_std""": [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], } SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , __A ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__A , __A ) def _snake_case ( self :Optional[Any] , **__A :Optional[int] ) -> Union[str, Any]: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , **__A ) def _snake_case ( self :Optional[int] , **__A :List[str] ) -> Union[str, Any]: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **__A ) def _snake_case ( self :str , **__A :Dict ) -> List[str]: """simple docstring""" return CLIPImageProcessor.from_pretrained(self.tmpdirname , **__A ) def _snake_case ( self :Union[str, Any] ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] SCREAMING_SNAKE_CASE__ = [Image.fromarray(np.moveaxis(__A , 0 , -1 ) ) for x in image_inputs] return image_inputs def _snake_case ( self :Dict ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) processor_slow.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=__A ) SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) processor_fast.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __A ) self.assertIsInstance(processor_fast.tokenizer , __A ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __A ) self.assertIsInstance(processor_fast.image_processor , __A ) def _snake_case ( self :Tuple ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE__ = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) SCREAMING_SNAKE_CASE__ = self.get_image_processor(do_normalize=__A , padding_value=1.0 ) SCREAMING_SNAKE_CASE__ = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __A ) def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = image_processor(__A , return_tensors="""np""" ) SCREAMING_SNAKE_CASE__ = processor(images=__A , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _snake_case ( self :Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) SCREAMING_SNAKE_CASE__ = """lower newer""" SCREAMING_SNAKE_CASE__ = processor(text=__A ) SCREAMING_SNAKE_CASE__ = tokenizer(__A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _snake_case ( self :Any ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) SCREAMING_SNAKE_CASE__ = """lower newer""" SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = processor(text=__A , images=__A ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__A ): processor() def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) SCREAMING_SNAKE_CASE__ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE__ = processor.batch_decode(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.batch_decode(__A ) self.assertListEqual(__A , __A ) def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_image_processor() SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = CLIPProcessor(tokenizer=__A , image_processor=__A ) SCREAMING_SNAKE_CASE__ = """lower newer""" SCREAMING_SNAKE_CASE__ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE__ = processor(text=__A , images=__A ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
6
"""simple docstring""" import socket def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase = socket.gethostname() UpperCAmelCase = 12312 sock.connect((host, port) ) sock.send(b"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: UpperCAmelCase = sock.recv(1024 ) if not data: break out_file.write(lowerCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
673
0
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class lowercase_ ( unittest.TestCase , __lowerCAmelCase ): '''simple docstring''' def lowerCAmelCase_ ( self : Optional[Any] ): _A = load_tool('text-classification' ) self.tool.setup() _A = load_tool('text-classification' , remote=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): _A = self.tool('That\'s quite cool' , ['positive', 'negative'] ) self.assertEqual(_UpperCAmelCase , 'positive' ) def lowerCAmelCase_ ( self : Union[str, Any] ): _A = self.remote_tool('That\'s quite cool' , ['positive', 'negative'] ) self.assertEqual(_UpperCAmelCase , 'positive' ) def lowerCAmelCase_ ( self : List[str] ): _A = self.tool(text='That\'s quite cool' , labels=['positive', 'negative'] ) self.assertEqual(_UpperCAmelCase , 'positive' ) def lowerCAmelCase_ ( self : Any ): _A = self.remote_tool(text='That\'s quite cool' , labels=['positive', 'negative'] ) self.assertEqual(_UpperCAmelCase , 'positive' )
7
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
0
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class SCREAMING_SNAKE_CASE (unittest.TestCase ): lowerCAmelCase = ViTImageProcessor if is_vision_available() else None @property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = (3, 32, 128) __A : List[str] = tempfile.mkdtemp() # fmt: off __A : Any = ['[GO]', '[s]', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] # fmt: on __A : Tuple = dict(zip(_UpperCAmelCase , range(len(_UpperCAmelCase)))) __A : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(_UpperCAmelCase) + '\n') __A : Tuple = { 'do_normalize': False, 'do_resize': True, 'image_processor_type': 'ViTImageProcessor', 'resample': 3, 'size': {'height': 32, 'width': 128}, } __A : Any = os.path.join(self.tmpdirname , _UpperCAmelCase) with open(self.image_processor_file , 'w' , encoding='utf-8') as fp: json.dump(_UpperCAmelCase , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , **_UpperCAmelCase): '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , **_UpperCAmelCase): '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' shutil.rmtree(self.tmpdirname) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[Any] = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta) __A : Optional[int] = Image.fromarray(np.moveaxis(_UpperCAmelCase , 0 , -1)) return image_input def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = self.get_tokenizer() __A : List[Any] = self.get_image_processor() __A : List[Any] = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) processor.save_pretrained(self.tmpdirname) __A : Any = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=_UpperCAmelCase) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.char_tokenizer , _UpperCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor.image_processor , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = self.get_tokenizer() __A : List[str] = self.get_image_processor() __A : int = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) processor.save_pretrained(self.tmpdirname) __A : int = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)') __A : Optional[Any] = self.get_image_processor(do_normalize=_UpperCAmelCase , padding_value=1.0) __A : Dict = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_UpperCAmelCase , padding_value=1.0) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.char_tokenizer , _UpperCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = self.get_image_processor() __A : Any = self.get_tokenizer() __A : Any = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : int = self.prepare_image_inputs() __A : Tuple = image_processor(_UpperCAmelCase , return_tensors='np') __A : Optional[Any] = processor(images=_UpperCAmelCase , return_tensors='np') for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = self.get_image_processor() __A : Any = self.get_tokenizer() __A : Union[str, Any] = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : Any = 'test' __A : Union[str, Any] = processor(text=_UpperCAmelCase) __A : Tuple = tokenizer(_UpperCAmelCase) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = self.get_image_processor() __A : Tuple = self.get_tokenizer() __A : List[str] = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : str = 'test' __A : Dict = self.prepare_image_inputs() __A : List[Any] = processor(text=_UpperCAmelCase , images=_UpperCAmelCase) self.assertListEqual(list(inputs.keys()) , ['pixel_values', 'labels']) # test if it raises when no input is passed with pytest.raises(_UpperCAmelCase): processor() def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[int] = self.get_image_processor() __A : int = self.get_tokenizer() __A : Optional[int] = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __A : str = processor.char_decode(_UpperCAmelCase) __A : Union[str, Any] = tokenizer.batch_decode(_UpperCAmelCase) __A : Optional[int] = [seq.replace(' ' , '') for seq in decoded_tok] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = self.get_image_processor() __A : Optional[Any] = self.get_tokenizer() __A : Dict = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : int = None __A : Any = self.prepare_image_inputs() __A : Union[str, Any] = processor(text=_UpperCAmelCase , images=_UpperCAmelCase) self.assertListEqual(list(inputs.keys()) , processor.model_input_names) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[int] = self.get_image_processor() __A : Optional[int] = self.get_tokenizer() __A : Optional[int] = MgpstrProcessor(tokenizer=_UpperCAmelCase , image_processor=_UpperCAmelCase) __A : Tuple = torch.randn(1 , 27 , 38) __A : Any = torch.randn(1 , 27 , 5_0257) __A : List[Any] = torch.randn(1 , 27 , 3_0522) __A : List[Any] = processor.batch_decode([char_input, bpe_input, wp_input]) self.assertListEqual(list(results.keys()) , ['generated_text', 'scores', 'char_preds', 'bpe_preds', 'wp_preds'])
8
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = list(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): UpperCAmelCase = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _lowerCAmelCase ( lowerCAmelCase = None , lowerCAmelCase = 128 ): '''simple docstring''' if function is None: return functools.partial(lowerCAmelCase , starting_batch_size=lowerCAmelCase ) UpperCAmelCase = starting_batch_size def decorator(*lowerCAmelCase , **lowerCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() UpperCAmelCase = list(inspect.signature(lowerCAmelCase ).parameters.keys() ) # Guard against user error if len(lowerCAmelCase ) < (len(lowerCAmelCase ) + 1): UpperCAmelCase = """, """.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) except Exception as e: if should_reduce_batch_size(lowerCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
673
0
import os import string import sys SCREAMING_SNAKE_CASE__ = 1 << 8 SCREAMING_SNAKE_CASE__ = { '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 2_7, '''up''': 6_5 + ARROW_KEY_FLAG, '''down''': 6_6 + ARROW_KEY_FLAG, '''right''': 6_7 + ARROW_KEY_FLAG, '''left''': 6_8 + ARROW_KEY_FLAG, '''mod_int''': 9_1, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 5_0, '''delete''': 5_1, '''pg_up''': 5_3, '''pg_down''': 5_4, } SCREAMING_SNAKE_CASE__ = KEYMAP['''up'''] SCREAMING_SNAKE_CASE__ = KEYMAP['''left'''] if sys.platform == "win32": SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = { b'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, b'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(1_0): SCREAMING_SNAKE_CASE__ = ord(str(i)) def A ( ) -> Union[str, Any]: if os.name == "nt": import msvcrt A__ = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(__UpperCamelCase ) == 0: # Read the keystroke A__ = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): A__ = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: A__ = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['mod_int'] ) ) WIN_CH_BUFFER.append(__UpperCamelCase ) if ord(__UpperCamelCase ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) A__ = chr(KEYMAP['esc'] ) except KeyError: A__ = cha[1] else: A__ = ch.decode(__UpperCamelCase ) else: A__ = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty A__ = sys.stdin.fileno() A__ = termios.tcgetattr(__UpperCamelCase ) try: tty.setraw(__UpperCamelCase ) A__ = sys.stdin.read(1 ) finally: termios.tcsetattr(__UpperCamelCase , termios.TCSADRAIN , __UpperCamelCase ) return ch def A ( ) -> Dict: A__ = get_raw_chars() if ord(__UpperCamelCase ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(__UpperCamelCase ) == KEYMAP["esc"]: A__ = get_raw_chars() if ord(__UpperCamelCase ) == KEYMAP["mod_int"]: A__ = get_raw_chars() if ord(__UpperCamelCase ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(__UpperCamelCase ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(__UpperCamelCase ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
9
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
0
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[vertex] def UpperCamelCase_ ( self : Optional[Any] , _A : List[str] , _A : Union[str, Any] ): _UpperCamelCase = pos def UpperCamelCase_ ( self : Any , _A : List[str] , _A : int , _A : Optional[Any] , _A : Union[str, Any] ): if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , _A ) self.top_to_bottom(_A , _A , _A , _A ) def UpperCamelCase_ ( self : List[str] , _A : Tuple , _A : Optional[Any] , _A : int , _A : Optional[int] ): _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , _A ) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(_A , _A ) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(_A , 0 ) def UpperCamelCase_ ( self : int , _A : Tuple , _A : int ): _UpperCamelCase = len(_A ) // 2 - 1 for i in range(_A , -1 , -1 ): self.top_to_bottom(_A , _A , len(_A ) , _A ) def UpperCamelCase_ ( self : Any , _A : int , _A : List[str] ): _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(_A , 0 , len(_A ) , _A ) return temp def _snake_case ( __snake_case ): _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case , __snake_case ) for _ in range(1 , len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case , __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case , heap.get_position(__snake_case ) , __snake_case , __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _lowerCAmelCase = int(input("Enter number of edges: ").strip()) _lowerCAmelCase = defaultdict(list) for _ in range(edges_number): _lowerCAmelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
10
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
0
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features lowercase_ = logging.get_logger(__name__) lowercase_ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) lowercase_ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __A : '''simple docstring''' __lowerCamelCase : str = field( default=A , metadata={'help': 'Model type selected in the list: ' + ', '.join(A )} ) __lowerCamelCase : str = field( default=A , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) __lowerCamelCase : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase : int = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) __lowerCamelCase : int = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) __lowerCamelCase : int = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) __lowerCamelCase : bool = field( default=A , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase : bool = field( default=A , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) __lowerCamelCase : float = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase : int = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) __lowerCamelCase : int = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) __lowerCamelCase : int = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class __A ( A ): '''simple docstring''' __lowerCamelCase : int = 'train' __lowerCamelCase : Union[str, Any] = 'dev' class __A ( A ): '''simple docstring''' __lowerCamelCase : SquadDataTrainingArguments __lowerCamelCase : List[SquadFeatures] __lowerCamelCase : Split __lowerCamelCase : bool def __init__(self , A , A , A = None , A = Split.train , A = False , A = None , A = "pt" , ) -> int: """simple docstring""" _a = args _a = is_language_sensitive _a = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(A , A ): try: _a = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) _a = mode # Load data features from cache or dataset file _a = '''v2''' if args.version_2_with_negative else '''v1''' _a = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _a = cached_features_file + '''.lock''' with FileLock(A ): if os.path.exists(A ) and not args.overwrite_cache: _a = time.time() _a = torch.load(A ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. _a = self.old_features['''features'''] _a = self.old_features.get('''dataset''' , A ) _a = self.old_features.get('''examples''' , A ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' ''' future run''' ) else: if mode == Split.dev: _a = self.processor.get_dev_examples(args.data_dir ) else: _a = self.processor.get_train_examples(args.data_dir ) _a , _a = squad_convert_examples_to_features( examples=self.examples , tokenizer=A , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=A , ) _a = time.time() torch.save( {'''features''': self.features, '''dataset''': self.dataset, '''examples''': self.examples} , A , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__(self ) -> List[Any]: """simple docstring""" return len(self.features ) def __getitem__(self , A ) -> Dict[str, torch.Tensor]: """simple docstring""" _a = self.features[i] _a = torch.tensor(feature.input_ids , dtype=torch.long ) _a = torch.tensor(feature.attention_mask , dtype=torch.long ) _a = torch.tensor(feature.token_type_ids , dtype=torch.long ) _a = torch.tensor(feature.cls_index , dtype=torch.long ) _a = torch.tensor(feature.p_mask , dtype=torch.float ) _a = torch.tensor(feature.is_impossible , dtype=torch.float ) _a = { '''input_ids''': input_ids, '''attention_mask''': attention_mask, '''token_type_ids''': token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({'''cls_index''': cls_index, '''p_mask''': p_mask} ) if self.args.version_2_with_negative: inputs.update({'''is_impossible''': is_impossible} ) if self.is_language_sensitive: inputs.update({'''langs''': (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: _a = torch.tensor(feature.start_position , dtype=torch.long ) _a = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({'''start_positions''': start_positions, '''end_positions''': end_positions} ) return inputs
11
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( a_ ): _A : Optional[int] = 'facebook/bart-large-mnli' _A : Union[str, Any] = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) _A : Dict = 'text_classifier' _A : Union[str, Any] = AutoTokenizer _A : Tuple = AutoModelForSequenceClassification _A : Optional[int] = ['text', ['text']] _A : Dict = ['text'] def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setup() UpperCAmelCase = self.model.config UpperCAmelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = labels return self.pre_processor( [text] * len(snake_case__ ) , [f'''This example is {label}''' for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def UpperCamelCase_ ( self , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = outputs.logits UpperCAmelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
673
0
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class _snake_case ( unittest.TestCase ): def lowercase__ ( self): '''simple docstring''' lowercase__ : Dict = """ylacombe/bark-small""" lowercase__ : Dict = tempfile.mkdtemp() lowercase__ : Any = """en_speaker_1""" lowercase__ : Optional[int] = """This is a test string""" lowercase__ : Tuple = """speaker_embeddings_path.json""" lowercase__ : str = """speaker_embeddings""" def lowercase__ ( self , **SCREAMING_SNAKE_CASE_): '''simple docstring''' return AutoTokenizer.from_pretrained(self.checkpoint , **SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowercase__ ( self): '''simple docstring''' lowercase__ : Dict = self.get_tokenizer() lowercase__ : int = BarkProcessor(tokenizer=SCREAMING_SNAKE_CASE_) processor.save_pretrained(self.tmpdirname) lowercase__ : List[str] = BarkProcessor.from_pretrained(self.tmpdirname) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab()) @slow def lowercase__ ( self): '''simple docstring''' lowercase__ : Union[str, Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) lowercase__ : Optional[int] = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""") lowercase__ : Any = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="""(BOS)""" , eos_token="""(EOS)""" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) def lowercase__ ( self): '''simple docstring''' lowercase__ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) lowercase__ : Optional[int] = 35 lowercase__ : Tuple = 2 lowercase__ : Dict = 8 lowercase__ : Optional[int] = { """semantic_prompt""": np.ones(SCREAMING_SNAKE_CASE_), """coarse_prompt""": np.ones((nb_codebooks_coarse, seq_len)), """fine_prompt""": np.ones((nb_codebooks_total, seq_len)), } # test providing already loaded voice_preset lowercase__ : Tuple = processor(text=self.input_string , voice_preset=SCREAMING_SNAKE_CASE_) lowercase__ : Union[str, Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(SCREAMING_SNAKE_CASE_ , np.array([])).tolist()) # test loading voice preset from npz file lowercase__ : List[Any] = os.path.join(self.tmpdirname , """file.npz""") np.savez(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_) lowercase__ : str = processor(text=self.input_string , voice_preset=SCREAMING_SNAKE_CASE_) lowercase__ : Optional[Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(SCREAMING_SNAKE_CASE_ , np.array([])).tolist()) # test loading voice preset from the hub lowercase__ : int = processor(text=self.input_string , voice_preset=self.voice_preset) def lowercase__ ( self): '''simple docstring''' lowercase__ : Dict = self.get_tokenizer() lowercase__ : str = BarkProcessor(tokenizer=SCREAMING_SNAKE_CASE_) lowercase__ : List[Any] = processor(text=self.input_string) lowercase__ : List[str] = tokenizer( self.input_string , padding="""max_length""" , max_length=2_56 , add_special_tokens=SCREAMING_SNAKE_CASE_ , return_attention_mask=SCREAMING_SNAKE_CASE_ , return_token_type_ids=SCREAMING_SNAKE_CASE_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist())
12
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase_ ( a_ ): _A : Union[List[PIL.Image.Image], np.ndarray] _A : Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
673
0
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) __lowerCamelCase : int = str(bin(UpperCAmelCase_ ) )[2:] # remove the leading "0b" __lowerCamelCase : Any = str(bin(UpperCAmelCase_ ) )[2:] # remove the leading "0b" __lowerCamelCase : List[str] = max(len(UpperCAmelCase_ ) , len(UpperCAmelCase_ ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(UpperCAmelCase_ ) , b_binary.zfill(UpperCAmelCase_ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
13
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : Any = { '''configuration_encodec''': [ '''ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EncodecConfig''', ], '''feature_extraction_encodec''': ['''EncodecFeatureExtractor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ '''ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EncodecModel''', '''EncodecPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys lowerCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
0
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def __UpperCAmelCase ( ) -> None: """simple docstring""" print('''Making key files...''' ) make_key_files('''rsa''' ,1_024 ) print('''Key files generation successful.''' ) def __UpperCAmelCase ( __a : int ) -> tuple[tuple[int, int], tuple[int, int]]: """simple docstring""" print('''Generating prime p...''' ) _a : Union[str, Any] = rabinMiller.generate_large_prime(__a ) print('''Generating prime q...''' ) _a : Optional[Any] = rabinMiller.generate_large_prime(__a ) _a : Union[str, Any] = p * q print('''Generating e that is relatively prime to (p - 1) * (q - 1)...''' ) while True: _a : Optional[int] = random.randrange(2 ** (key_size - 1) ,2 ** (key_size) ) if cryptoMath.gcd(__a ,(p - 1) * (q - 1) ) == 1: break print('''Calculating d that is mod inverse of e...''' ) _a : Optional[int] = cryptoMath.find_mod_inverse(__a ,(p - 1) * (q - 1) ) _a : Optional[Any] = (n, e) _a : int = (n, d) return (public_key, private_key) def __UpperCAmelCase ( __a : str ,__a : int ) -> None: """simple docstring""" if os.path.exists(F"""{name}_pubkey.txt""" ) or os.path.exists(F"""{name}_privkey.txt""" ): print('''\nWARNING:''' ) print( F"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" '''Use a different name or delete these files and re-run this program.''' ) sys.exit() _a , _a : List[str] = generate_key(__a ) print(F"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(F"""{name}_pubkey.txt""" ,'''w''' ) as out_file: out_file.write(F"""{key_size},{public_key[0]},{public_key[1]}""" ) print(F"""Writing private key to file {name}_privkey.txt...""" ) with open(F"""{name}_privkey.txt""" ,'''w''' ) as out_file: out_file.write(F"""{key_size},{private_key[0]},{private_key[1]}""" ) if __name__ == "__main__": main()
14
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
673
0
def UpperCamelCase ( __magic_name__ : int , __magic_name__ : int ) -> str: """simple docstring""" if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) lowercase__ = str(bin(__magic_name__ ) )[2:] # remove the leading "0b" lowercase__ = str(bin(__magic_name__ ) )[2:] # remove the leading "0b" lowercase__ = max(len(__magic_name__ ) , len(__magic_name__ ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(__magic_name__ ) , b_binary.zfill(__magic_name__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
15
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = VideoToVideoSDPipeline _A : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} _A : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} _A : int = PipelineTesterMixin.required_optional_params - {'latents'} _A : List[str] = False # No `output_type`. _A : Any = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) UpperCAmelCase = CLIPTextModel(snake_case__ ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase_ ( self , snake_case__ , snake_case__=0 ) -> List[str]: """simple docstring""" UpperCAmelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(snake_case__ ) else: UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = VideoToVideoSDPipeline(**snake_case__ ) UpperCAmelCase = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) UpperCAmelCase = """np""" UpperCAmelCase = sd_pipe(**snake_case__ ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case__ ) UpperCAmelCase = video.to("""cuda""" ) UpperCAmelCase = """Spiderman is surfing""" UpperCAmelCase = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type="""pt""" ).frames UpperCAmelCase = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
673
0
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self : List[str] ): SCREAMING_SNAKE_CASE = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) SCREAMING_SNAKE_CASE = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) SCREAMING_SNAKE_CASE = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def _snake_case ( self : str ): print(f"Found {torch.cuda.device_count()} devices." ) SCREAMING_SNAKE_CASE = ["torchrun", f"--nproc_per_node={torch.cuda.device_count()}", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__lowerCamelCase , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self : List[Any] ): print(f"Found {torch.cuda.device_count()} devices." ) SCREAMING_SNAKE_CASE = ["torchrun", f"--nproc_per_node={torch.cuda.device_count()}", self.operation_file_path] print(f"Command: {cmd}" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__lowerCamelCase , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self : Optional[int] ): SCREAMING_SNAKE_CASE = ["torchrun", f"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__lowerCamelCase , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self : Union[str, Any] ): print(f"Found {torch.cuda.device_count()} devices, using 2 devices only" ) SCREAMING_SNAKE_CASE = ["torchrun", f"--nproc_per_node={torch.cuda.device_count()}", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(__lowerCamelCase , env=os.environ.copy() ) if __name__ == "__main__": __A : Tuple = Accelerator() __A : Optional[Any] = (accelerator.state.process_index + 2, 1_0) __A : Union[str, Any] = torch.randint(0, 1_0, shape).to(accelerator.device) __A : Optional[Any] = '' __A : int = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." __A : Optional[int] = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." __A : str = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
16
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class UpperCamelCase_ ( a_ ): _A : int = 'wav2vec2' def __init__( self , snake_case__=32 , snake_case__=7_68 , snake_case__=12 , snake_case__=12 , snake_case__=30_72 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=1e-5 , snake_case__="group" , snake_case__="gelu" , snake_case__=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case__=(5, 2, 2, 2, 2, 2, 2) , snake_case__=(10, 3, 3, 3, 3, 2, 2) , snake_case__=False , snake_case__=1_28 , snake_case__=16 , snake_case__=False , snake_case__=True , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__=3_20 , snake_case__=2 , snake_case__=0.1 , snake_case__=1_00 , snake_case__=2_56 , snake_case__=2_56 , snake_case__=0.1 , snake_case__="sum" , snake_case__=False , snake_case__=False , snake_case__=2_56 , snake_case__=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case__=(5, 3, 3, 1, 1) , snake_case__=(1, 2, 3, 1, 1) , snake_case__=5_12 , snake_case__=0 , snake_case__=1 , snake_case__=2 , snake_case__=False , snake_case__=3 , snake_case__=2 , snake_case__=3 , snake_case__=None , snake_case__=None , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) UpperCAmelCase = hidden_size UpperCAmelCase = feat_extract_norm UpperCAmelCase = feat_extract_activation UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = conv_bias UpperCAmelCase = num_conv_pos_embeddings UpperCAmelCase = num_conv_pos_embedding_groups UpperCAmelCase = len(self.conv_dim ) UpperCAmelCase = num_hidden_layers UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = feat_proj_dropout UpperCAmelCase = final_dropout UpperCAmelCase = layerdrop UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = vocab_size UpperCAmelCase = do_stable_layer_norm UpperCAmelCase = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase = apply_spec_augment UpperCAmelCase = mask_time_prob UpperCAmelCase = mask_time_length UpperCAmelCase = mask_time_min_masks UpperCAmelCase = mask_feature_prob UpperCAmelCase = mask_feature_length UpperCAmelCase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCAmelCase = num_codevectors_per_group UpperCAmelCase = num_codevector_groups UpperCAmelCase = contrastive_logits_temperature UpperCAmelCase = feat_quantizer_dropout UpperCAmelCase = num_negatives UpperCAmelCase = codevector_dim UpperCAmelCase = proj_codevector_dim UpperCAmelCase = diversity_loss_weight # ctc loss UpperCAmelCase = ctc_loss_reduction UpperCAmelCase = ctc_zero_infinity # adapter UpperCAmelCase = add_adapter UpperCAmelCase = adapter_kernel_size UpperCAmelCase = adapter_stride UpperCAmelCase = num_adapter_layers UpperCAmelCase = output_hidden_size or hidden_size UpperCAmelCase = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = xvector_output_dim @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
673
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase_ : Any = { '''configuration_data2vec_audio''': ['''DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecAudioConfig'''], '''configuration_data2vec_text''': [ '''DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecTextConfig''', '''Data2VecTextOnnxConfig''', ], '''configuration_data2vec_vision''': [ '''DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecVisionConfig''', '''Data2VecVisionOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : List[str] = [ '''DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecAudioForAudioFrameClassification''', '''Data2VecAudioForCTC''', '''Data2VecAudioForSequenceClassification''', '''Data2VecAudioForXVector''', '''Data2VecAudioModel''', '''Data2VecAudioPreTrainedModel''', ] UpperCAmelCase_ : List[Any] = [ '''DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecTextForCausalLM''', '''Data2VecTextForMaskedLM''', '''Data2VecTextForMultipleChoice''', '''Data2VecTextForQuestionAnswering''', '''Data2VecTextForSequenceClassification''', '''Data2VecTextForTokenClassification''', '''Data2VecTextModel''', '''Data2VecTextPreTrainedModel''', ] UpperCAmelCase_ : List[str] = [ '''DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecVisionForImageClassification''', '''Data2VecVisionForMaskedImageModeling''', '''Data2VecVisionForSemanticSegmentation''', '''Data2VecVisionModel''', '''Data2VecVisionPreTrainedModel''', ] if is_tf_available(): UpperCAmelCase_ : int = [ '''TFData2VecVisionForImageClassification''', '''TFData2VecVisionForSemanticSegmentation''', '''TFData2VecVisionModel''', '''TFData2VecVisionPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
17
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCAmelCase_ : Optional[Any] = NewType('''DataClass''', Any) lowerCAmelCase_ : Any = NewType('''DataClassType''', Any) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = {str(lowerCAmelCase ): choice for choice in choices} return lambda lowerCAmelCase : str_to_choice.get(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( *, lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = None , **lowerCAmelCase , ): '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCAmelCase = {} if aliases is not None: UpperCAmelCase = aliases if help is not None: UpperCAmelCase = help return dataclasses.field(metadata=lowerCAmelCase , default=lowerCAmelCase , default_factory=lowerCAmelCase , **lowerCAmelCase ) class UpperCamelCase_ ( a_ ): _A : Iterable[DataClassType] def __init__( self , snake_case__ , **snake_case__ ) -> List[str]: """simple docstring""" if "formatter_class" not in kwargs: UpperCAmelCase = ArgumentDefaultsHelpFormatter super().__init__(**snake_case__ ) if dataclasses.is_dataclass(snake_case__ ): UpperCAmelCase = [dataclass_types] UpperCAmelCase = list(snake_case__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(snake_case__ ) @staticmethod def UpperCamelCase_ ( snake_case__ , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = f'''--{field.name}''' UpperCAmelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , snake_case__ ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) UpperCAmelCase = kwargs.pop("""aliases""" , [] ) if isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [aliases] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(snake_case__ , """UnionType""" ) and isinstance(snake_case__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(snake_case__ ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" f''' Problem encountered in field \'{field.name}\'.''' ) if type(snake_case__ ) not in field.type.__args__: # filter `str` in Union UpperCAmelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCAmelCase = ( field.type.__args__[0] if isinstance(snake_case__ , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCAmelCase = {} if origin_type is Literal or (isinstance(field.type , snake_case__ ) and issubclass(field.type , snake_case__ )): if origin_type is Literal: UpperCAmelCase = field.type.__args__ else: UpperCAmelCase = [x.value for x in field.type] UpperCAmelCase = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default else: UpperCAmelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCAmelCase = copy(snake_case__ ) # Hack because type=bool in argparse does not behave as we want. UpperCAmelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCAmelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCAmelCase = default # This tells argparse we accept 0 or 1 value after --field_name UpperCAmelCase = """?""" # This is the value that will get picked if we do --field_name (without value) UpperCAmelCase = True elif isclass(snake_case__ ) and issubclass(snake_case__ , snake_case__ ): UpperCAmelCase = field.type.__args__[0] UpperCAmelCase = """+""" if field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() elif field.default is dataclasses.MISSING: UpperCAmelCase = True else: UpperCAmelCase = field.type if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default elif field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() else: UpperCAmelCase = True parser.add_argument(snake_case__ , *snake_case__ , **snake_case__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCAmelCase = False parser.add_argument(f'''--no_{field.name}''' , action="""store_false""" , dest=field.name , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Any: """simple docstring""" if hasattr(snake_case__ , """_argument_group_name""" ): UpperCAmelCase = self.add_argument_group(dtype._argument_group_name ) else: UpperCAmelCase = self try: UpperCAmelCase = get_type_hints(snake_case__ ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(snake_case__ ): UpperCAmelCase = """.""".join(map(snake_case__ , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(snake_case__ ): if not field.init: continue UpperCAmelCase = type_hints[field.name] self._parse_dataclass_field(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self , snake_case__=None , snake_case__=False , snake_case__=True , snake_case__=None , snake_case__=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCAmelCase = [] if args_filename: args_files.append(Path(snake_case__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCAmelCase = ArgumentParser() args_file_parser.add_argument(snake_case__ , type=snake_case__ , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCAmelCase , UpperCAmelCase = args_file_parser.parse_known_args(args=snake_case__ ) UpperCAmelCase = vars(snake_case__ ).get(args_file_flag.lstrip("""-""" ) , snake_case__ ) if cmd_args_file_paths: args_files.extend([Path(snake_case__ ) for p in cmd_args_file_paths] ) UpperCAmelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCAmelCase = file_args + args if args is not None else file_args + sys.argv[1:] UpperCAmelCase , UpperCAmelCase = self.parse_known_args(args=snake_case__ ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in vars(snake_case__ ).items() if k in keys} for k in keys: delattr(snake_case__ , snake_case__ ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(snake_case__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = set(args.keys() ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(snake_case__ )}''' ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(snake_case__ ) , encoding="""utf-8""" ) as open_json_file: UpperCAmelCase = json.loads(open_json_file.read() ) UpperCAmelCase = self.parse_dict(snake_case__ , allow_extra_keys=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = self.parse_dict(yaml.safe_load(Path(snake_case__ ).read_text() ) , allow_extra_keys=snake_case__ ) return tuple(snake_case__ )
673
0
'''simple docstring''' from __future__ import annotations class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ) -> None: _lowerCAmelCase = data _lowerCAmelCase = None _lowerCAmelCase = None def __a(SCREAMING_SNAKE_CASE_ : Node | None ): # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def __a(SCREAMING_SNAKE_CASE_ : Node | None ): '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def __a(SCREAMING_SNAKE_CASE_ : Node ): '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def __a(): # Main function for testing. '''simple docstring''' _lowerCAmelCase = Node(1 ) _lowerCAmelCase = Node(2 ) _lowerCAmelCase = Node(3 ) _lowerCAmelCase = Node(4 ) _lowerCAmelCase = Node(5 ) _lowerCAmelCase = Node(6 ) _lowerCAmelCase = Node(7 ) _lowerCAmelCase = Node(8 ) _lowerCAmelCase = Node(9 ) print(is_full_binary_tree(SCREAMING_SNAKE_CASE_ ) ) print(depth_of_tree(SCREAMING_SNAKE_CASE_ ) ) print("Tree is: " ) display(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
18
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCAmelCase_ : List[str] = False class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self , snake_case__=32 ) -> Optional[Any]: """simple docstring""" set_seed(0 ) UpperCAmelCase = UNetaDModel(sample_size=snake_case__ , in_channels=3 , out_channels=3 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) UpperCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randint(0 , 10_00 , (4,) ).long().to(snake_case__ ) for _ in range(4 )] # train with a DDPM scheduler UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) )
673
0
"""simple docstring""" import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _a = logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" warnings.warn( '''The preprocess method is deprecated and will be removed in a future version. Please''' ''' use VaeImageProcessor.preprocess instead''', __snake_case, ) if isinstance(__snake_case, torch.Tensor ): return image elif isinstance(__snake_case, PIL.Image.Image ): _UpperCamelCase = [image] if isinstance(image[0], PIL.Image.Image ): _UpperCamelCase , _UpperCamelCase = image[0].size _UpperCamelCase , _UpperCamelCase = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 _UpperCamelCase = [np.array(i.resize((w, h), resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] _UpperCamelCase = np.concatenate(__snake_case, axis=0 ) _UpperCamelCase = np.array(__snake_case ).astype(np.floataa ) / 255.0 _UpperCamelCase = image.transpose(0, 3, 1, 2 ) _UpperCamelCase = 2.0 * image - 1.0 _UpperCamelCase = torch.from_numpy(__snake_case ) elif isinstance(image[0], torch.Tensor ): _UpperCamelCase = torch.cat(__snake_case, dim=0 ) return image def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if isinstance(__snake_case, torch.Tensor ): return mask elif isinstance(__snake_case, PIL.Image.Image ): _UpperCamelCase = [mask] if isinstance(mask[0], PIL.Image.Image ): _UpperCamelCase , _UpperCamelCase = mask[0].size _UpperCamelCase , _UpperCamelCase = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 _UpperCamelCase = [np.array(m.convert('''L''' ).resize((w, h), resample=PIL_INTERPOLATION['''nearest'''] ) )[None, :] for m in mask] _UpperCamelCase = np.concatenate(__snake_case, axis=0 ) _UpperCamelCase = mask.astype(np.floataa ) / 255.0 _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = torch.from_numpy(__snake_case ) elif isinstance(mask[0], torch.Tensor ): _UpperCamelCase = torch.cat(__snake_case, dim=0 ) return mask class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 42 lowercase__ = 42 def __init__( self , __a , __a) -> List[str]: '''simple docstring''' super().__init__() self.register_modules(unet=__a , scheduler=__a) @torch.no_grad() def __call__( self , __a , __a , __a = 2_50 , __a = 0.0 , __a = 10 , __a = 10 , __a = None , __a = "pil" , __a = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' _UpperCamelCase = image _UpperCamelCase = _preprocess_image(__a) _UpperCamelCase = original_image.to(device=self.device , dtype=self.unet.dtype) _UpperCamelCase = _preprocess_mask(__a) _UpperCamelCase = mask_image.to(device=self.device , dtype=self.unet.dtype) _UpperCamelCase = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__a , __a) and len(__a) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(__a)}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''') _UpperCamelCase = original_image.shape _UpperCamelCase = randn_tensor(__a , generator=__a , device=self.device , dtype=self.unet.dtype) # set step values self.scheduler.set_timesteps(__a , __a , __a , self.device) _UpperCamelCase = eta _UpperCamelCase = self.scheduler.timesteps[0] + 1 _UpperCamelCase = generator[0] if isinstance(__a , __a) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps)): if t < t_last: # predict the noise residual _UpperCamelCase = self.unet(__a , __a).sample # compute previous image: x_t -> x_t-1 _UpperCamelCase = self.scheduler.step(__a , __a , __a , __a , __a , __a).prev_sample else: # compute the reverse: x_t-1 -> x_t _UpperCamelCase = self.scheduler.undo_step(__a , __a , __a) _UpperCamelCase = t _UpperCamelCase = (image / 2 + 0.5).clamp(0 , 1) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": _UpperCamelCase = self.numpy_to_pil(__a) if not return_dict: return (image,) return ImagePipelineOutput(images=__a)
19
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class UpperCamelCase_ : def __init__( self , snake_case__=2 , snake_case__=3 , snake_case__=64 , snake_case__=None ) -> List[str]: """simple docstring""" UpperCAmelCase = np.random.default_rng(snake_case__ ) UpperCAmelCase = length UpperCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) UpperCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> int: """simple docstring""" return self.length def __getitem__( self , snake_case__ ) -> Tuple: """simple docstring""" return {"x": self.x[i], "y": self.y[i]} class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[str]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> List[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a[0] + self.b[0] class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> Optional[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a + self.b def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = 16 ): '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer UpperCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) UpperCAmelCase = {"""train""": """tests/test_samples/MRPC/train.csv""", """validation""": """tests/test_samples/MRPC/dev.csv"""} UpperCAmelCase = load_dataset("""csv""" , data_files=lowerCAmelCase ) UpperCAmelCase = datasets["""train"""].unique("""label""" ) UpperCAmelCase = {v: i for i, v in enumerate(lowerCAmelCase )} def tokenize_function(lowerCAmelCase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase = tokenizer( examples["""sentence1"""] , examples["""sentence2"""] , truncation=lowerCAmelCase , max_length=lowerCAmelCase , padding="""max_length""" ) if "label" in examples: UpperCAmelCase = [label_to_id[l] for l in examples["""label"""]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase = datasets.map( lowerCAmelCase , batched=lowerCAmelCase , remove_columns=["""sentence1""", """sentence2""", """label"""] , ) def collate_fn(lowerCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return tokenizer.pad(lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. UpperCAmelCase = DataLoader(tokenized_datasets["""train"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=2 ) UpperCAmelCase = DataLoader(tokenized_datasets["""validation"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
673
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase: Union[str, Any] = { 'configuration_roformer': ['ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoFormerConfig', 'RoFormerOnnxConfig'], 'tokenization_roformer': ['RoFormerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Any = ['RoFormerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Tuple = [ 'ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoFormerForCausalLM', 'RoFormerForMaskedLM', 'RoFormerForMultipleChoice', 'RoFormerForQuestionAnswering', 'RoFormerForSequenceClassification', 'RoFormerForTokenClassification', 'RoFormerLayer', 'RoFormerModel', 'RoFormerPreTrainedModel', 'load_tf_weights_in_roformer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Optional[int] = [ 'TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRoFormerForCausalLM', 'TFRoFormerForMaskedLM', 'TFRoFormerForMultipleChoice', 'TFRoFormerForQuestionAnswering', 'TFRoFormerForSequenceClassification', 'TFRoFormerForTokenClassification', 'TFRoFormerLayer', 'TFRoFormerModel', 'TFRoFormerPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: int = [ 'FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxRoFormerForMaskedLM', 'FlaxRoFormerForMultipleChoice', 'FlaxRoFormerForQuestionAnswering', 'FlaxRoFormerForSequenceClassification', 'FlaxRoFormerForTokenClassification', 'FlaxRoFormerModel', 'FlaxRoFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys _lowerCAmelCase: Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
20
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = hidden_states.shape UpperCAmelCase = jax.image.resize( snake_case__ , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Any: """simple docstring""" UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : int = None _A : float = 0.0 _A : bool = None _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.in_channels if self.out_channels is None else self.out_channels UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = nn.Dense(snake_case__ , dtype=self.dtype ) UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Dropout(self.dropout_prob ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut UpperCAmelCase = None if use_nin_shortcut: UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self , snake_case__ , snake_case__ , snake_case__=True ) -> List[Any]: """simple docstring""" UpperCAmelCase = hidden_states UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) UpperCAmelCase = self.time_emb_proj(nn.swish(snake_case__ ) ) UpperCAmelCase = jnp.expand_dims(jnp.expand_dims(snake_case__ , 1 ) , 1 ) UpperCAmelCase = hidden_states + temb UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.dropout(snake_case__ , snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) if self.conv_shortcut is not None: UpperCAmelCase = self.conv_shortcut(snake_case__ ) return hidden_states + residual
673
0
from __future__ import annotations def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): __magic_name__ : Tuple =sorted(numsa + numsa ) __magic_name__ , __magic_name__ : Optional[Any] =divmod(len(lowerCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ : Union[str, Any] = [float(x) for x in input("Enter the elements of first array: ").split()] UpperCAmelCase_ : Dict = [float(x) for x in input("Enter the elements of second array: ").split()] print(F"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
21
"""simple docstring""" from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=30 , snake_case__=2 , snake_case__=3 , snake_case__=True , snake_case__=True , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=10 , snake_case__=0.02 , snake_case__=3 , snake_case__=None , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 1 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModel(config=snake_case__ ) UpperCAmelCase = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) UpperCAmelCase = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[int] = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _A : Optional[Any] = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) _A : Optional[int] = False _A : Any = False _A : List[str] = False def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(snake_case__ ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=snake_case__ , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**snake_case__ ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case__ ) UpperCAmelCase = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , snake_case__ , atol=1e-4 )
673
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : Union[str, Any] = logging.get_logger(__name__) _snake_case : Optional[int] = { 'asapp/sew-d-tiny-100k': 'https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class A ( _a ): lowercase_ = 'sew-d' def __init__( self : Tuple , lowerCAmelCase_ : Dict=32 , lowerCAmelCase_ : str=7_68 , lowerCAmelCase_ : List[str]=12 , lowerCAmelCase_ : Tuple=12 , lowerCAmelCase_ : Tuple=30_72 , lowerCAmelCase_ : str=2 , lowerCAmelCase_ : Dict=5_12 , lowerCAmelCase_ : List[str]=2_56 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=("p2c", "c2p") , lowerCAmelCase_ : Dict="layer_norm" , lowerCAmelCase_ : List[Any]="gelu_python" , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=0.0 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : Dict=0.0_2 , lowerCAmelCase_ : Any=1e-7 , lowerCAmelCase_ : Any=1e-5 , lowerCAmelCase_ : Tuple="group" , lowerCAmelCase_ : Union[str, Any]="gelu" , lowerCAmelCase_ : Optional[int]=(64, 1_28, 1_28, 1_28, 1_28, 2_56, 2_56, 2_56, 2_56, 5_12, 5_12, 5_12, 5_12) , lowerCAmelCase_ : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowerCAmelCase_ : Tuple=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : Union[str, Any]=1_28 , lowerCAmelCase_ : Union[str, Any]=16 , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : Any=0.0_5 , lowerCAmelCase_ : Optional[int]=10 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Optional[Any]=0.0 , lowerCAmelCase_ : Tuple=10 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple="mean" , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : List[str]=False , lowerCAmelCase_ : Tuple=2_56 , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : List[str]=1 , lowerCAmelCase_ : Any=2 , **lowerCAmelCase_ : str , ) -> Dict: """simple docstring""" super().__init__(**lowerCAmelCase_ , pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ ) _a = hidden_size _a = feat_extract_norm _a = feat_extract_activation _a = list(lowerCAmelCase_ ) _a = list(lowerCAmelCase_ ) _a = list(lowerCAmelCase_ ) _a = conv_bias _a = num_conv_pos_embeddings _a = num_conv_pos_embedding_groups _a = len(self.conv_dim ) _a = num_hidden_layers _a = intermediate_size _a = squeeze_factor _a = max_position_embeddings _a = position_buckets _a = share_att_key _a = relative_attention _a = norm_rel_ebd _a = list(lowerCAmelCase_ ) _a = hidden_act _a = num_attention_heads _a = hidden_dropout _a = attention_dropout _a = activation_dropout _a = feat_proj_dropout _a = final_dropout _a = layer_norm_eps _a = feature_layer_norm_eps _a = initializer_range _a = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F'but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)' F'= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _a = apply_spec_augment _a = mask_time_prob _a = mask_time_length _a = mask_time_min_masks _a = mask_feature_prob _a = mask_feature_length _a = mask_feature_min_masks # ctc loss _a = ctc_loss_reduction _a = ctc_zero_infinity # sequence classification _a = use_weighted_layer_sum _a = classifier_proj_size @property def __lowerCAmelCase ( self : int ) -> List[str]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
22
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=5_12 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = NystromformerForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = NystromformerForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = self.num_choices UpperCAmelCase = NystromformerForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[Any] = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) _A : Optional[Any] = ( { 'feature-extraction': NystromformerModel, 'fill-mask': NystromformerForMaskedLM, 'question-answering': NystromformerForQuestionAnswering, 'text-classification': NystromformerForSequenceClassification, 'token-classification': NystromformerForTokenClassification, 'zero-shot': NystromformerForSequenceClassification, } if is_torch_available() else {} ) _A : int = False _A : Dict = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = NystromformerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case__ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = NystromformerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): UpperCAmelCase = model(snake_case__ )[0] UpperCAmelCase = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """the [MASK] of Belgium is Brussels""" UpperCAmelCase = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = tokenizer(snake_case__ , return_tensors="""pt""" ) with torch.no_grad(): UpperCAmelCase = model(encoding.input_ids ).logits UpperCAmelCase = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(snake_case__ ) , """capital""" )
673
0
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType snake_case__ : List[str] = logging.get_logger(__name__) class _a ( UpperCAmelCase__ ): """simple docstring""" A_ = """vision-encoder-decoder""" A_ = True def __init__( self , **_UpperCAmelCase ) -> Dict: super().__init__(**_UpperCAmelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"""A configuraton of type {self.model_type} cannot be instantiated because """ f"""not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}""" ) UpperCamelCase_ = kwargs.pop('encoder' ) UpperCamelCase_ = encoder_config.pop('model_type' ) UpperCamelCase_ = kwargs.pop('decoder' ) UpperCamelCase_ = decoder_config.pop('model_type' ) UpperCamelCase_ = AutoConfig.for_model(_UpperCAmelCase , **_UpperCAmelCase ) UpperCamelCase_ = AutoConfig.for_model(_UpperCAmelCase , **_UpperCAmelCase ) UpperCamelCase_ = True @classmethod def _UpperCAmelCase ( cls , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ) -> PretrainedConfig: logger.info('Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) UpperCamelCase_ = True UpperCamelCase_ = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_UpperCAmelCase ) def _UpperCAmelCase ( self ) -> int: UpperCamelCase_ = copy.deepcopy(self.__dict__ ) UpperCamelCase_ = self.encoder.to_dict() UpperCamelCase_ = self.decoder.to_dict() UpperCamelCase_ = self.__class__.model_type return output class _a ( UpperCAmelCase__ ): """simple docstring""" A_ = version.parse("""1.11""" ) @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4 @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({'last_hidden_state': {0: 'batch', 1: 'encoder_sequence'}} ) class _a ( UpperCAmelCase__ ): """simple docstring""" @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: UpperCamelCase_ = OrderedDict() UpperCamelCase_ = {0: 'batch', 1: 'past_decoder_sequence + sequence'} UpperCamelCase_ = {0: 'batch', 1: 'past_decoder_sequence + sequence'} UpperCamelCase_ = {0: 'batch', 1: 'encoder_sequence'} return common_inputs def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase = -1 , _UpperCAmelCase = -1 , _UpperCAmelCase = False , _UpperCAmelCase = None , ) -> Mapping[str, Any]: import torch UpperCamelCase_ = OrderedDict() UpperCamelCase_ = super().generate_dummy_inputs( _UpperCAmelCase , batch_size=_UpperCAmelCase , seq_length=_UpperCAmelCase , is_pair=_UpperCAmelCase , framework=_UpperCAmelCase ) UpperCamelCase_ , UpperCamelCase_ = dummy_input['input_ids'].shape UpperCamelCase_ = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCamelCase_ = dummy_input.pop('input_ids' ) UpperCamelCase_ = dummy_input.pop('attention_mask' ) UpperCamelCase_ = torch.zeros(_UpperCAmelCase ) return common_inputs class _a ( UpperCAmelCase__ ): """simple docstring""" @property def _UpperCAmelCase ( self ) -> None: pass def _UpperCAmelCase ( self , _UpperCAmelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_UpperCAmelCase ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = "default" ) -> OnnxConfig: UpperCamelCase_ = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_UpperCAmelCase , _UpperCAmelCase )
23
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Optional[int] = False def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return TrainCommand(lowerCAmelCase ) class UpperCamelCase_ ( a_ ): @staticmethod def UpperCamelCase_ ( snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = parser.add_parser("""train""" , help="""CLI tool to train a model on a task.""" ) train_parser.add_argument( """--train_data""" , type=snake_case__ , required=snake_case__ , help="""path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.""" , ) train_parser.add_argument( """--column_label""" , type=snake_case__ , default=0 , help="""Column of the dataset csv file with example labels.""" ) train_parser.add_argument( """--column_text""" , type=snake_case__ , default=1 , help="""Column of the dataset csv file with example texts.""" ) train_parser.add_argument( """--column_id""" , type=snake_case__ , default=2 , help="""Column of the dataset csv file with example ids.""" ) train_parser.add_argument( """--skip_first_row""" , action="""store_true""" , help="""Skip the first row of the csv file (headers).""" ) train_parser.add_argument("""--validation_data""" , type=snake_case__ , default="""""" , help="""path to validation dataset.""" ) train_parser.add_argument( """--validation_split""" , type=snake_case__ , default=0.1 , help="""if validation dataset is not provided, fraction of train dataset to use as validation dataset.""" , ) train_parser.add_argument("""--output""" , type=snake_case__ , default="""./""" , help="""path to saved the trained model.""" ) train_parser.add_argument( """--task""" , type=snake_case__ , default="""text_classification""" , help="""Task to train the model on.""" ) train_parser.add_argument( """--model""" , type=snake_case__ , default="""bert-base-uncased""" , help="""Model's name or path to stored model.""" ) train_parser.add_argument("""--train_batch_size""" , type=snake_case__ , default=32 , help="""Batch size for training.""" ) train_parser.add_argument("""--valid_batch_size""" , type=snake_case__ , default=64 , help="""Batch size for validation.""" ) train_parser.add_argument("""--learning_rate""" , type=snake_case__ , default=3e-5 , help="""Learning rate.""" ) train_parser.add_argument("""--adam_epsilon""" , type=snake_case__ , default=1e-08 , help="""Epsilon for Adam optimizer.""" ) train_parser.set_defaults(func=snake_case__ ) def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = logging.get_logger("""transformers-cli/training""" ) UpperCAmelCase = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=snake_case__ ) UpperCAmelCase = args.output UpperCAmelCase = args.column_label UpperCAmelCase = args.column_text UpperCAmelCase = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": UpperCAmelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = args.validation_split UpperCAmelCase = args.train_batch_size UpperCAmelCase = args.valid_batch_size UpperCAmelCase = args.learning_rate UpperCAmelCase = args.adam_epsilon def UpperCamelCase_ ( self ) -> Any: """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" raise NotImplementedError def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
673
0
'''simple docstring''' import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase): __lowercase : List[Any] = VQModel __lowercase : str = '''sample''' @property def lowerCAmelCase ( self , __SCREAMING_SNAKE_CASE=(32, 32) ) -> Dict: '''simple docstring''' __snake_case = 4 __snake_case = 3 __snake_case = floats_tensor((batch_size, num_channels) + sizes ).to(__SCREAMING_SNAKE_CASE ) return {"sample": image} @property def lowerCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' return (3, 32, 32) @property def lowerCAmelCase ( self ) -> Tuple: '''simple docstring''' return (3, 32, 32) def lowerCAmelCase ( self ) -> Optional[int]: '''simple docstring''' __snake_case = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 3, } __snake_case = self.dummy_input return init_dict, inputs_dict def lowerCAmelCase ( self ) -> Tuple: '''simple docstring''' pass def lowerCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' pass def lowerCAmelCase ( self ) -> Tuple: '''simple docstring''' __snake_case , __snake_case = VQModel.from_pretrained('''fusing/vqgan-dummy''' , output_loading_info=__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(__SCREAMING_SNAKE_CASE ) __snake_case = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def lowerCAmelCase ( self ) -> Any: '''simple docstring''' __snake_case = VQModel.from_pretrained('''fusing/vqgan-dummy''' ) model.to(__SCREAMING_SNAKE_CASE ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) __snake_case = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) __snake_case = image.to(__SCREAMING_SNAKE_CASE ) with torch.no_grad(): __snake_case = model(__SCREAMING_SNAKE_CASE ).sample __snake_case = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off __snake_case = torch.tensor([-0.0_153, -0.4_044, -0.1_880, -0.5_161, -0.2_418, -0.4_072, -0.1_612, -0.0_633, -0.0_143] ) # fmt: on self.assertTrue(torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) )
24
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=sys.maxsize ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """bilinear""" UpperCAmelCase = max_size UpperCAmelCase = short_edge_length def __call__( self , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = [] for img in imgs: UpperCAmelCase , UpperCAmelCase = img.shape[:2] # later: provide list and randomly choose index for resize UpperCAmelCase = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img UpperCAmelCase = size * 1.0 / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size if max(snake_case__ , snake_case__ ) > self.max_size: UpperCAmelCase = self.max_size * 1.0 / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase = int(neww + 0.5 ) UpperCAmelCase = int(newh + 0.5 ) if img.dtype == np.uinta: UpperCAmelCase = Image.fromarray(snake_case__ ) UpperCAmelCase = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) UpperCAmelCase = np.asarray(snake_case__ ) else: UpperCAmelCase = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw UpperCAmelCase = nn.functional.interpolate( snake_case__ , (newh, neww) , mode=self.interp_method , align_corners=snake_case__ ).squeeze(0 ) img_augs.append(snake_case__ ) return img_augs class UpperCamelCase_ : def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) UpperCAmelCase = cfg.INPUT.FORMAT UpperCAmelCase = cfg.SIZE_DIVISIBILITY UpperCAmelCase = cfg.PAD_VALUE UpperCAmelCase = cfg.INPUT.MAX_SIZE_TEST UpperCAmelCase = cfg.MODEL.DEVICE UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = lambda snake_case__ : (x - self.pixel_mean) / self.pixel_std def UpperCamelCase_ ( self , snake_case__ ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = tuple(max(snake_case__ ) for s in zip(*[img.shape for img in images] ) ) UpperCAmelCase = [im.shape[-2:] for im in images] UpperCAmelCase = [ nn.functional.pad( snake_case__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(snake_case__ , snake_case__ ) ] return torch.stack(snake_case__ ), torch.tensor(snake_case__ ) def __call__( self , snake_case__ , snake_case__=False ) -> Optional[Any]: """simple docstring""" with torch.no_grad(): if not isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [images] if single_image: assert len(snake_case__ ) == 1 for i in range(len(snake_case__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(snake_case__ , images.pop(snake_case__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( snake_case__ , torch.as_tensor(img_tensorize(images.pop(snake_case__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge UpperCAmelCase = torch.tensor([im.shape[:2] for im in images] ) UpperCAmelCase = self.aug(snake_case__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic UpperCAmelCase = [self.normalizer(snake_case__ ) for x in images] # now pad them to do the following operations UpperCAmelCase , UpperCAmelCase = self.pad(snake_case__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad UpperCAmelCase = torch.true_divide(snake_case__ , snake_case__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' assert torch.isfinite(lowerCAmelCase ).all(), "Box tensor contains infinite or NaN!" UpperCAmelCase , UpperCAmelCase = box_size tensor[:, 0].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 1].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 2].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 3].clamp_(min=0 , max=lowerCAmelCase )
673
0
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : Optional[Any] = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(_a , _a) def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Union[str, Any] = emb.weight.shape SCREAMING_SNAKE_CASE : int = nn.Linear(_a , _a , bias=_a) SCREAMING_SNAKE_CASE : str = emb.weight.data return lin_layer def lowerCamelCase__ ( _a , _a="facebook/mbart-large-en-ro" , _a=False , _a=False): SCREAMING_SNAKE_CASE : List[str] = torch.load(_a , map_location="cpu")["model"] remove_ignore_keys_(_a) SCREAMING_SNAKE_CASE : List[Any] = state_dict["encoder.embed_tokens.weight"].shape[0] SCREAMING_SNAKE_CASE : Optional[int] = MBartConfig.from_pretrained(_a , vocab_size=_a) if mbart_aa and finetuned: SCREAMING_SNAKE_CASE : int = "relu" SCREAMING_SNAKE_CASE : Optional[Any] = state_dict["decoder.embed_tokens.weight"] SCREAMING_SNAKE_CASE : Optional[Any] = MBartForConditionalGeneration(_a) model.model.load_state_dict(_a) if finetuned: SCREAMING_SNAKE_CASE : str = make_linear_from_emb(model.model.shared) return model if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') a_ = parser.parse_args() a_ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
25
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : List[str] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase = """""" else: UpperCAmelCase = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = dct.pop(lowerCAmelCase ) UpperCAmelCase = val def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase = 1000 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = int(deit_name[-6:-4] ) UpperCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): UpperCAmelCase = 192 UpperCAmelCase = 768 UpperCAmelCase = 12 UpperCAmelCase = 3 elif deit_name[9:].startswith("""small""" ): UpperCAmelCase = 384 UpperCAmelCase = 1536 UpperCAmelCase = 12 UpperCAmelCase = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 # load original model from timm UpperCAmelCase = timm.create_model(lowerCAmelCase , pretrained=lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase = timm_model.state_dict() UpperCAmelCase = create_rename_keys(lowerCAmelCase , lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) read_in_q_k_v(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model UpperCAmelCase = DeiTForImageClassificationWithTeacher(lowerCAmelCase ).eval() model.load_state_dict(lowerCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase = DeiTImageProcessor(size=lowerCAmelCase , crop_size=config.image_size ) UpperCAmelCase = image_processor(images=prepare_img() , return_tensors="""pt""" ) UpperCAmelCase = encoding["""pixel_values"""] UpperCAmelCase = model(lowerCAmelCase ) UpperCAmelCase = timm_model(lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase_ : str = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
673
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging __UpperCamelCase = logging.get_logger(__name__) class _A ( __lowercase ): lowercase__: Tuple = ['''input_values''', '''padding_mask'''] def __init__( self : List[Any] , __magic_name__ : int = 1 , __magic_name__ : int = 2_40_00 , __magic_name__ : float = 0.0 , __magic_name__ : float = None , __magic_name__ : float = None , **__magic_name__ : Dict , ) -> Dict: """simple docstring""" super().__init__(feature_size=__magic_name__ , sampling_rate=__magic_name__ , padding_value=__magic_name__ , **__magic_name__ ) __snake_case : Dict = chunk_length_s __snake_case : Optional[Any] = overlap @property def lowercase__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def lowercase__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self : Union[str, Any] , __magic_name__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __magic_name__ : Optional[Union[bool, str, PaddingStrategy]] = None , __magic_name__ : Optional[bool] = False , __magic_name__ : Optional[int] = None , __magic_name__ : Optional[Union[str, TensorType]] = None , __magic_name__ : Optional[int] = None , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) if padding and truncation: raise ValueError("""Both padding and truncation were set. Make sure you only set one.""" ) elif padding is None: # by default let's pad the inputs __snake_case : Optional[Any] = True __snake_case : Any = bool( isinstance(__magic_name__ , (list, tuple) ) and (isinstance(raw_audio[0] , (np.ndarray, tuple, list) )) ) if is_batched: __snake_case : Tuple = [np.asarray(__magic_name__ , dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(__magic_name__ , np.ndarray ): __snake_case : Any = np.asarray(__magic_name__ , dtype=np.floataa ) elif isinstance(__magic_name__ , np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): __snake_case : List[Any] = raw_audio.astype(np.floataa ) # always return batch if not is_batched: __snake_case : str = [np.asarray(__magic_name__ ).T] # verify inputs are valid for idx, example in enumerate(__magic_name__ ): if example.ndim > 2: raise ValueError(f'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(f'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(f'''Expected stereo audio but example has {example.shape[-1]} channels''' ) __snake_case : Dict = None __snake_case : Union[str, Any] = BatchFeature({"""input_values""": raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: __snake_case : Optional[int] = min(array.shape[0] for array in raw_audio ) __snake_case : Union[str, Any] = int(np.floor(max_length / self.chunk_stride ) ) __snake_case : int = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: __snake_case : Dict = max(array.shape[0] for array in raw_audio ) __snake_case : str = int(np.ceil(max_length / self.chunk_stride ) ) __snake_case : List[str] = (nb_step - 1) * self.chunk_stride + self.chunk_length __snake_case : Tuple = """max_length""" else: __snake_case : Dict = input_values # normal padding on batch if padded_inputs is None: __snake_case : str = self.pad( __magic_name__ , max_length=__magic_name__ , truncation=__magic_name__ , padding=__magic_name__ , return_attention_mask=__magic_name__ , ) if padding: __snake_case : List[Any] = padded_inputs.pop("""attention_mask""" ) __snake_case : Tuple = [] for example in padded_inputs.pop("""input_values""" ): if self.feature_size == 1: __snake_case : Optional[Any] = example[..., None] input_values.append(example.T ) __snake_case : List[Any] = input_values if return_tensors is not None: __snake_case : List[Any] = padded_inputs.convert_to_tensors(__magic_name__ ) return padded_inputs
26
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = 32 , snake_case__ = True , snake_case__ = 1 / 2_55 , snake_case__ = True , snake_case__ = True , snake_case__ = [0.48_145_466, 0.4_578_275, 0.40_821_073] , snake_case__ = [0.26_862_954, 0.26_130_258, 0.27_577_711] , snake_case__ = True , snake_case__=7 , snake_case__=30 , snake_case__=4_00 , snake_case__=3 , ) -> List[str]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = do_resize UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_88} UpperCAmelCase = size_divisor UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_pad UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase_ ( self , snake_case__ , snake_case__=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = image_inputs[0] if isinstance(snake_case__ , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] UpperCAmelCase = size / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size UpperCAmelCase = int((13_33 / 8_00) * size ) if max(snake_case__ , snake_case__ ) > max_size: UpperCAmelCase = max_size / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase , UpperCAmelCase = int(newh + 0.5 ), int(neww + 0.5 ) UpperCAmelCase , UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """size_divisor""" ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
673
0
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __A : Union[str, Any] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class lowerCamelCase( nn.Module ): '''simple docstring''' def __init__( self , snake_case_ ): super().__init__() _A = torchvision.models.resnetaaa(pretrained=snake_case_ ) _A = list(model.children() )[:-2] _A = nn.Sequential(*snake_case_ ) _A = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def lowerCAmelCase__ ( self , snake_case_ ): # Bx3x224x224 -> Bx2048x7x7 -> Bx2048xN -> BxNx2048 _A = self.pool(self.model(snake_case_ ) ) _A = torch.flatten(snake_case_ , start_dim=2 ) _A = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class lowerCamelCase( __snake_case ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _A = [json.loads(snake_case_ ) for l in open(snake_case_ )] _A = os.path.dirname(snake_case_ ) _A = tokenizer _A = labels _A = len(snake_case_ ) _A = max_seq_length _A = transforms def __len__( self ): return len(self.data ) def __getitem__( self , snake_case_ ): _A = torch.LongTensor(self.tokenizer.encode(self.data[index]['text'] , add_special_tokens=snake_case_ ) ) _A, _A, _A = sentence[0], sentence[1:-1], sentence[-1] _A = sentence[: self.max_seq_length] _A = torch.zeros(self.n_classes ) _A = 1 _A = Image.open(os.path.join(self.data_dir , self.data[index]['img'] ) ).convert('RGB' ) _A = self.transforms(snake_case_ ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def lowerCAmelCase__ ( self ): _A = Counter() for row in self.data: label_freqs.update(row['label'] ) return label_freqs def __lowerCAmelCase( _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" _A = [len(row['sentence'] ) for row in batch] _A, _A = len(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _A = torch.zeros(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , dtype=torch.long ) _A = torch.zeros(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ): _A = input_row['sentence'] _A = 1 _A = torch.stack([row['image'] for row in batch] ) _A = torch.stack([row['label'] for row in batch] ) _A = torch.stack([row['image_start_token'] for row in batch] ) _A = torch.stack([row['image_end_token'] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def __lowerCAmelCase( ) -> Dict: """simple docstring""" return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def __lowerCAmelCase( ) -> str: """simple docstring""" return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.4677_7044, 0.4453_1429, 0.4066_1017] , std=[0.1222_1994, 0.1214_5835, 0.1438_0469] , ), ] )
27
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[str] = XLMRobertaTokenizer _A : List[str] = XLMRobertaTokenizerFast _A : Optional[Any] = True _A : List[str] = True def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(snake_case__ ) , 10_02 ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(snake_case__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) UpperCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=snake_case__ ) UpperCAmelCase = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = """I was born in 92000, and this is falsé.""" UpperCAmelCase = tokenizer.tokenize(snake_case__ ) UpperCAmelCase = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
673
0
'''simple docstring''' import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput UpperCamelCase_ = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self, *A, A=None, A=None, A=None, **A ): '''simple docstring''' super().__init__(*A, **A ) SCREAMING_SNAKE_CASE : Optional[Any] = eval_examples SCREAMING_SNAKE_CASE : List[Any] = post_process_function SCREAMING_SNAKE_CASE : List[Any] = quant_trainer_args SCREAMING_SNAKE_CASE : Tuple = 128 # default number of calibration samples def UpperCamelCase_ ( self, A=None ): '''simple docstring''' if calib_dataset is None and self.calib_dataset is None: raise ValueError('Trainer: calibration requires an calib_dataset.' ) SCREAMING_SNAKE_CASE : List[str] = calib_dataset if calib_dataset is not None else self.calib_dataset SCREAMING_SNAKE_CASE : Tuple = self._remove_unused_columns(A, description='Calibration' ) return DataLoader( A, batch_size=self.args.eval_batch_size, collate_fn=self.data_collator, drop_last=self.args.dataloader_drop_last, num_workers=self.args.dataloader_num_workers, pin_memory=self.args.dataloader_pin_memory, shuffle=A, ) def UpperCamelCase_ ( self, A=None ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.train_dataset if calib_dataset is None else calib_dataset SCREAMING_SNAKE_CASE : Any = self.get_calib_dataloader(A ) SCREAMING_SNAKE_CASE : List[Any] = self.model quant_trainer.configure_model(A, self.quant_trainer_args, calib=A ) model.eval() quant_trainer.enable_calibration(A ) logger.info('***** Running calibration *****' ) logger.info(F" Num examples = {self.calib_num}" ) logger.info(F" Batch size = {calib_dataloader.batch_size}" ) for step, inputs in enumerate(A ): # Prediction step SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = self.prediction_step(A, A, prediction_loss_only=A ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(A, self.quant_trainer_args ) SCREAMING_SNAKE_CASE : List[Any] = model def UpperCamelCase_ ( self, A=None, A=None, A=None, A = "eval" ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.eval_dataset if eval_dataset is None else eval_dataset SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_eval_dataloader(A ) SCREAMING_SNAKE_CASE : Dict = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : int = self.compute_metrics SCREAMING_SNAKE_CASE : Optional[int] = None SCREAMING_SNAKE_CASE : List[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : List[str] = eval_loop( A, description='Evaluation', prediction_loss_only=True if compute_metrics is None else None, ignore_keys=A, ) finally: SCREAMING_SNAKE_CASE : List[str] = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: SCREAMING_SNAKE_CASE : Tuple = self.post_process_function(A, A, output.predictions ) SCREAMING_SNAKE_CASE : Optional[int] = self.compute_metrics(A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"{metric_key_prefix}_" ): SCREAMING_SNAKE_CASE : List[Any] = metrics.pop(A ) self.log(A ) else: SCREAMING_SNAKE_CASE : Optional[int] = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) SCREAMING_SNAKE_CASE : Dict = self.callback_handler.on_evaluate(self.args, self.state, self.control, A ) return metrics def UpperCamelCase_ ( self, A, A, A=None, A = "test" ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = self.get_test_dataloader(A ) # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : List[str] = self.compute_metrics SCREAMING_SNAKE_CASE : str = None SCREAMING_SNAKE_CASE : Dict = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : int = eval_loop( A, description='Prediction', prediction_loss_only=True if compute_metrics is None else None, ignore_keys=A, ) finally: SCREAMING_SNAKE_CASE : Tuple = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output SCREAMING_SNAKE_CASE : int = self.post_process_function(A, A, output.predictions, 'predict' ) SCREAMING_SNAKE_CASE : Union[str, Any] = self.compute_metrics(A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"{metric_key_prefix}_" ): SCREAMING_SNAKE_CASE : Optional[int] = metrics.pop(A ) return PredictionOutput(predictions=predictions.predictions, label_ids=predictions.label_ids, metrics=A ) def UpperCamelCase_ ( self, A="./" ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = self.eval_dataset SCREAMING_SNAKE_CASE : int = self.get_eval_dataloader(A ) SCREAMING_SNAKE_CASE : Union[str, Any] = next(iter(A ) ) # saving device - to make it consistent SCREAMING_SNAKE_CASE : Optional[Any] = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) # convert to tuple SCREAMING_SNAKE_CASE : Optional[int] = tuple(v.to(A ) for k, v in batch.items() ) logger.info('Converting model to be onnx compatible' ) from pytorch_quantization.nn import TensorQuantizer SCREAMING_SNAKE_CASE : List[str] = True SCREAMING_SNAKE_CASE : Dict = self.model.to(A ) model.eval() model.float() SCREAMING_SNAKE_CASE : int = model.module if hasattr(A, 'module' ) else model quant_trainer.configure_model(A, self.quant_trainer_args ) SCREAMING_SNAKE_CASE : Any = os.path.join(A, 'model.onnx' ) logger.info(F"exporting model to {output_model_file}" ) SCREAMING_SNAKE_CASE : List[str] = {0: 'batch_size', 1: 'seq_len'} torch.onnx.export( A, A, A, export_params=A, opset_version=13, do_constant_folding=A, input_names=['input_ids', 'attention_mask', 'token_type_ids'], output_names=['output_start_logits', 'output_end_logits'], dynamic_axes={ 'input_ids': axes, 'attention_mask': axes, 'token_type_ids': axes, 'output_start_logits': axes, 'output_end_logits': axes, }, verbose=A, ) logger.info('onnx export finished' )
28
"""simple docstring""" import socket def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase = socket.gethostname() UpperCAmelCase = 12312 sock.connect((host, port) ) sock.send(b"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: UpperCAmelCase = sock.recv(1024 ) if not data: break out_file.write(lowerCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
673
0
"""simple docstring""" import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingface_hub.utils import ( EntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, is_jinja_available, ) from packaging import version from requests import HTTPError from .. import __version__ from .constants import ( DEPRECATED_REVISION_ARGS, DIFFUSERS_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, SAFETENSORS_WEIGHTS_NAME, WEIGHTS_NAME, ) from .import_utils import ( ENV_VARS_TRUE_VALUES, _flax_version, _jax_version, _onnxruntime_version, _torch_version, is_flax_available, is_onnx_available, is_torch_available, ) from .logging import get_logger A_ = get_logger(__name__) A_ = Path(__file__).parent / """model_card_template.md""" A_ = uuida().hex A_ = os.getenv("""HF_HUB_OFFLINE""", """""").upper() in ENV_VARS_TRUE_VALUES A_ = os.getenv("""DISABLE_TELEMETRY""", """""").upper() in ENV_VARS_TRUE_VALUES A_ = HUGGINGFACE_CO_RESOLVE_ENDPOINT + """/api/telemetry/""" def lowercase ( lowerCAmelCase__ = None ): lowerCamelCase_ = f"diffusers/{__version__}; python/{sys.version.split()[0]}; session_id/{SESSION_ID}" if DISABLE_TELEMETRY or HF_HUB_OFFLINE: return ua + "; telemetry/off" if is_torch_available(): ua += f"; torch/{_torch_version}" if is_flax_available(): ua += f"; jax/{_jax_version}" ua += f"; flax/{_flax_version}" if is_onnx_available(): ua += f"; onnxruntime/{_onnxruntime_version}" # CI will set this value to True if os.environ.get('''DIFFUSERS_IS_CI''' ,'''''' ).upper() in ENV_VARS_TRUE_VALUES: ua += "; is_ci/true" if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): ua += "; " + "; ".join(f"{k}/{v}" for k, v in user_agent.items() ) elif isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): ua += "; " + user_agent return ua def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ = None ,lowerCAmelCase__ = None ): if token is None: lowerCamelCase_ = HfFolder.get_token() if organization is None: lowerCamelCase_ = whoami(lowerCAmelCase__ )['''name'''] return f"{username}/{model_id}" else: return f"{organization}/{model_id}" def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ ): if not is_jinja_available(): raise ValueError( '''Modelcard rendering is based on Jinja templates.''' ''' Please make sure to have `jinja` installed before using `create_model_card`.''' ''' To install it, please run `pip install Jinja2`.''' ) if hasattr(lowerCAmelCase__ ,'''local_rank''' ) and args.local_rank not in [-1, 0]: return lowerCamelCase_ = args.hub_token if hasattr(lowerCAmelCase__ ,'''hub_token''' ) else None lowerCamelCase_ = get_full_repo_name(lowerCAmelCase__ ,token=lowerCAmelCase__ ) lowerCamelCase_ = ModelCard.from_template( card_data=ModelCardData( # Card metadata object that will be converted to YAML block language='''en''' ,license='''apache-2.0''' ,library_name='''diffusers''' ,tags=[] ,datasets=args.dataset_name ,metrics=[] ,) ,template_path=lowerCAmelCase__ ,model_name=lowerCAmelCase__ ,repo_name=lowerCAmelCase__ ,dataset_name=args.dataset_name if hasattr(lowerCAmelCase__ ,'''dataset_name''' ) else None ,learning_rate=args.learning_rate ,train_batch_size=args.train_batch_size ,eval_batch_size=args.eval_batch_size ,gradient_accumulation_steps=( args.gradient_accumulation_steps if hasattr(lowerCAmelCase__ ,'''gradient_accumulation_steps''' ) else None ) ,adam_betaa=args.adam_betaa if hasattr(lowerCAmelCase__ ,'''adam_beta1''' ) else None ,adam_betaa=args.adam_betaa if hasattr(lowerCAmelCase__ ,'''adam_beta2''' ) else None ,adam_weight_decay=args.adam_weight_decay if hasattr(lowerCAmelCase__ ,'''adam_weight_decay''' ) else None ,adam_epsilon=args.adam_epsilon if hasattr(lowerCAmelCase__ ,'''adam_epsilon''' ) else None ,lr_scheduler=args.lr_scheduler if hasattr(lowerCAmelCase__ ,'''lr_scheduler''' ) else None ,lr_warmup_steps=args.lr_warmup_steps if hasattr(lowerCAmelCase__ ,'''lr_warmup_steps''' ) else None ,ema_inv_gamma=args.ema_inv_gamma if hasattr(lowerCAmelCase__ ,'''ema_inv_gamma''' ) else None ,ema_power=args.ema_power if hasattr(lowerCAmelCase__ ,'''ema_power''' ) else None ,ema_max_decay=args.ema_max_decay if hasattr(lowerCAmelCase__ ,'''ema_max_decay''' ) else None ,mixed_precision=args.mixed_precision ,) lowerCamelCase_ = os.path.join(args.output_dir ,'''README.md''' ) model_card.save(lowerCAmelCase__ ) def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ = None ): if resolved_file is None or commit_hash is not None: return commit_hash lowerCamelCase_ = str(Path(lowerCAmelCase__ ).as_posix() ) lowerCamelCase_ = re.search(R'''snapshots/([^/]+)/''' ,lowerCAmelCase__ ) if search is None: return None lowerCamelCase_ = search.groups()[0] return commit_hash if REGEX_COMMIT_HASH.match(lowerCAmelCase__ ) else None # Old default cache path, potentially to be migrated. # This logic was more or less taken from `transformers`, with the following differences: # - Diffusers doesn't use custom environment variables to specify the cache path. # - There is no need to migrate the cache format, just move the files to the new location. A_ = os.path.expanduser( os.getenv("""HF_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """huggingface""")) ) A_ = os.path.join(hf_cache_home, """diffusers""") def lowercase ( lowerCAmelCase__ = None ,lowerCAmelCase__ = None ): if new_cache_dir is None: lowerCamelCase_ = DIFFUSERS_CACHE if old_cache_dir is None: lowerCamelCase_ = old_diffusers_cache lowerCamelCase_ = Path(lowerCAmelCase__ ).expanduser() lowerCamelCase_ = Path(lowerCAmelCase__ ).expanduser() for old_blob_path in old_cache_dir.glob('''**/blobs/*''' ): if old_blob_path.is_file() and not old_blob_path.is_symlink(): lowerCamelCase_ = new_cache_dir / old_blob_path.relative_to(lowerCAmelCase__ ) new_blob_path.parent.mkdir(parents=lowerCAmelCase__ ,exist_ok=lowerCAmelCase__ ) os.replace(lowerCAmelCase__ ,lowerCAmelCase__ ) try: os.symlink(lowerCAmelCase__ ,lowerCAmelCase__ ) except OSError: logger.warning( '''Could not create symlink between old cache and new cache. If you use an older version of diffusers again, files will be re-downloaded.''' ) # At this point, old_cache_dir contains symlinks to the new cache (it can still be used). A_ = os.path.join(DIFFUSERS_CACHE, """version_diffusers_cache.txt""") if not os.path.isfile(cache_version_file): A_ = 0 else: with open(cache_version_file) as f: try: A_ = int(f.read()) except ValueError: A_ = 0 if cache_version < 1: A_ = os.path.isdir(old_diffusers_cache) and len(os.listdir(old_diffusers_cache)) > 0 if old_cache_is_not_empty: logger.warning( """The cache for model files in Diffusers v0.14.0 has moved to a new location. Moving your """ """existing cached models. This is a one-time operation, you can interrupt it or run it """ """later by calling `diffusers.utils.hub_utils.move_cache()`.""" ) try: move_cache() except Exception as e: A_ = """\n""".join(traceback.format_tb(e.__traceback__)) logger.error( f"There was a problem when trying to move your cache:\n\n{trace}\n{e.__class__.__name__}: {e}\n\nPlease " """file an issue at https://github.com/huggingface/diffusers/issues/new/choose, copy paste this whole """ """message and we will do our best to help.""" ) if cache_version < 1: try: os.makedirs(DIFFUSERS_CACHE, exist_ok=True) with open(cache_version_file, """w""") as f: f.write("""1""") except Exception: logger.warning( f"There was a problem when trying to write in your cache folder ({DIFFUSERS_CACHE}). Please, ensure " """the directory exists and can be written to.""" ) def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ = None ): if variant is not None: lowerCamelCase_ = weights_name.split('''.''' ) lowerCamelCase_ = splits[:-1] + [variant] + splits[-1:] lowerCamelCase_ = '''.'''.join(lowerCAmelCase__ ) return weights_name def lowercase ( lowerCAmelCase__ ,*, lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__=None ,): lowerCamelCase_ = str(lowerCAmelCase__ ) if os.path.isfile(lowerCAmelCase__ ): return pretrained_model_name_or_path elif os.path.isdir(lowerCAmelCase__ ): if os.path.isfile(os.path.join(lowerCAmelCase__ ,lowerCAmelCase__ ) ): # Load from a PyTorch checkpoint lowerCamelCase_ = os.path.join(lowerCAmelCase__ ,lowerCAmelCase__ ) return model_file elif subfolder is not None and os.path.isfile( os.path.join(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) ): lowerCamelCase_ = os.path.join(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) return model_file else: raise EnvironmentError( f"Error no file named {weights_name} found in directory {pretrained_model_name_or_path}." ) else: # 1. First check if deprecated way of loading from branches is used if ( revision in DEPRECATED_REVISION_ARGS and (weights_name == WEIGHTS_NAME or weights_name == SAFETENSORS_WEIGHTS_NAME) and version.parse(version.parse(lowerCAmelCase__ ).base_version ) >= version.parse('''0.20.0''' ) ): try: lowerCamelCase_ = hf_hub_download( lowerCAmelCase__ ,filename=_add_variant(lowerCAmelCase__ ,lowerCAmelCase__ ) ,cache_dir=lowerCAmelCase__ ,force_download=lowerCAmelCase__ ,proxies=lowerCAmelCase__ ,resume_download=lowerCAmelCase__ ,local_files_only=lowerCAmelCase__ ,use_auth_token=lowerCAmelCase__ ,user_agent=lowerCAmelCase__ ,subfolder=lowerCAmelCase__ ,revision=revision or commit_hash ,) warnings.warn( f"Loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'` is deprecated. Loading instead from `revision='main'` with `variant={revision}`. Loading model variants via `revision='{revision}'` will be removed in diffusers v1. Please use `variant='{revision}'` instead." ,lowerCAmelCase__ ,) return model_file except: # noqa: E722 warnings.warn( f"You are loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'`. This behavior is deprecated and will be removed in diffusers v1. One should use `variant='{revision}'` instead. However, it appears that {pretrained_model_name_or_path} currently does not have a {_add_variant(lowerCAmelCase__ ,lowerCAmelCase__ )} file in the 'main' branch of {pretrained_model_name_or_path}. \n The Diffusers team and community would be very grateful if you could open an issue: https://github.com/huggingface/diffusers/issues/new with the title '{pretrained_model_name_or_path} is missing {_add_variant(lowerCAmelCase__ ,lowerCAmelCase__ )}' so that the correct variant file can be added." ,lowerCAmelCase__ ,) try: # 2. Load model file as usual lowerCamelCase_ = hf_hub_download( lowerCAmelCase__ ,filename=lowerCAmelCase__ ,cache_dir=lowerCAmelCase__ ,force_download=lowerCAmelCase__ ,proxies=lowerCAmelCase__ ,resume_download=lowerCAmelCase__ ,local_files_only=lowerCAmelCase__ ,use_auth_token=lowerCAmelCase__ ,user_agent=lowerCAmelCase__ ,subfolder=lowerCAmelCase__ ,revision=revision or commit_hash ,) return model_file except RepositoryNotFoundError: raise EnvironmentError( f"{pretrained_model_name_or_path} is not a local folder and is not a valid model identifier " '''listed on \'https://huggingface.co/models\'\nIf this is a private repository, make sure to pass a ''' '''token having permission to this repo with `use_auth_token` or log in with `huggingface-cli ''' '''login`.''' ) except RevisionNotFoundError: raise EnvironmentError( f"{revision} is not a valid git identifier (branch name, tag name or commit id) that exists for " '''this model name. Check the model page at ''' f"'https://huggingface.co/{pretrained_model_name_or_path}' for available revisions." ) except EntryNotFoundError: raise EnvironmentError( f"{pretrained_model_name_or_path} does not appear to have a file named {weights_name}." ) except HTTPError as err: raise EnvironmentError( f"There was a specific connection error when trying to load {pretrained_model_name_or_path}:\n{err}" ) except ValueError: raise EnvironmentError( f"We couldn't connect to '{HUGGINGFACE_CO_RESOLVE_ENDPOINT}' to load this model, couldn't find it" f" in the cached files and it looks like {pretrained_model_name_or_path} is not the path to a" f" directory containing a file named {weights_name} or" ''' \nCheckout your internet connection or see how to run the library in''' ''' offline mode at \'https://huggingface.co/docs/diffusers/installation#offline-mode\'.''' ) except EnvironmentError: raise EnvironmentError( f"Can't load the model for '{pretrained_model_name_or_path}'. If you were trying to load it from " '''\'https://huggingface.co/models\', make sure you don\'t have a local directory with the same name. ''' f"Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a directory " f"containing a file named {weights_name}" )
29
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
0
import colorsys from PIL import Image # type: ignore def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = x UpperCAmelCase_ : List[str] = y for step in range(_lowercase ): # noqa: B007 UpperCAmelCase_ : Union[str, Any] = a * a - b * b + x UpperCAmelCase_ : List[Any] = 2 * a * b + y UpperCAmelCase_ : List[Any] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def lowerCamelCase__ ( _lowercase ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def lowerCamelCase__ ( _lowercase ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255 ) for i in colorsys.hsv_to_rgb(_lowercase , 1 , 1 ) ) def lowerCamelCase__ ( _lowercase = 800 , _lowercase = 600 , _lowercase = -0.6 , _lowercase = 0 , _lowercase = 3.2 , _lowercase = 50 , _lowercase = True , ): '''simple docstring''' UpperCAmelCase_ : int = Image.new('''RGB''' , (image_width, image_height) ) UpperCAmelCase_ : List[str] = img.load() # loop through the image-coordinates for image_x in range(_lowercase ): for image_y in range(_lowercase ): # determine the figure-coordinates based on the image-coordinates UpperCAmelCase_ : str = figure_width / image_width * image_height UpperCAmelCase_ : Dict = figure_center_x + (image_x / image_width - 0.5) * figure_width UpperCAmelCase_ : Any = figure_center_y + (image_y / image_height - 0.5) * figure_height UpperCAmelCase_ : int = get_distance(_lowercase , _lowercase , _lowercase ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: UpperCAmelCase_ : Dict = get_color_coded_rgb(_lowercase ) else: UpperCAmelCase_ : int = get_black_and_white_rgb(_lowercase ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure __a = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
30
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = list(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): UpperCAmelCase = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _lowerCAmelCase ( lowerCAmelCase = None , lowerCAmelCase = 128 ): '''simple docstring''' if function is None: return functools.partial(lowerCAmelCase , starting_batch_size=lowerCAmelCase ) UpperCAmelCase = starting_batch_size def decorator(*lowerCAmelCase , **lowerCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() UpperCAmelCase = list(inspect.signature(lowerCAmelCase ).parameters.keys() ) # Guard against user error if len(lowerCAmelCase ) < (len(lowerCAmelCase ) + 1): UpperCAmelCase = """, """.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) except Exception as e: if should_reduce_batch_size(lowerCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
673
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ : str = logging.get_logger(__name__) lowerCamelCase__ : Tuple = { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/config.json', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/config.json', 'funnel-transformer/medium-base': 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json', 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/config.json', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json', 'funnel-transformer/xlarge-base': 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json', } class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "funnel" lowercase_ = { "hidden_size": "d_model", "num_attention_heads": "n_head", } def __init__( self : int , _lowerCAmelCase : Optional[int]=30_522 , _lowerCAmelCase : List[str]=[4, 4, 4] , _lowerCAmelCase : Tuple=None , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : int=768 , _lowerCAmelCase : Optional[Any]=12 , _lowerCAmelCase : Optional[Any]=64 , _lowerCAmelCase : Optional[Any]=3_072 , _lowerCAmelCase : List[str]="gelu_new" , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : int=0.1 , _lowerCAmelCase : Tuple=0.0 , _lowerCAmelCase : List[Any]=0.1 , _lowerCAmelCase : Dict=None , _lowerCAmelCase : str=1E-9 , _lowerCAmelCase : Any="mean" , _lowerCAmelCase : Union[str, Any]="relative_shift" , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Tuple=True , **_lowerCAmelCase : Optional[Any] , ): SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = block_sizes SCREAMING_SNAKE_CASE_ = [1] * len(_lowerCAmelCase ) if block_repeats is None else block_repeats assert len(_lowerCAmelCase ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." SCREAMING_SNAKE_CASE_ = num_decoder_layers SCREAMING_SNAKE_CASE_ = d_model SCREAMING_SNAKE_CASE_ = n_head SCREAMING_SNAKE_CASE_ = d_head SCREAMING_SNAKE_CASE_ = d_inner SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout SCREAMING_SNAKE_CASE_ = attention_dropout SCREAMING_SNAKE_CASE_ = activation_dropout SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = initializer_std SCREAMING_SNAKE_CASE_ = layer_norm_eps assert pooling_type in [ "mean", "max", ], F"Got {pooling_type} for `pooling_type` but only 'mean' and 'max' are supported." SCREAMING_SNAKE_CASE_ = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F"Got {attention_type} for `attention_type` but only 'relative_shift' and 'factorized' are supported." SCREAMING_SNAKE_CASE_ = attention_type SCREAMING_SNAKE_CASE_ = separate_cls SCREAMING_SNAKE_CASE_ = truncate_seq SCREAMING_SNAKE_CASE_ = pool_q_only super().__init__(**_lowerCAmelCase ) @property def lowerCAmelCase_ ( self : Optional[int] ): return sum(self.block_sizes ) @num_hidden_layers.setter def lowerCAmelCase_ ( self : int , _lowerCAmelCase : List[Any] ): raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.' ) @property def lowerCAmelCase_ ( self : List[Any] ): return len(self.block_sizes ) @num_blocks.setter def lowerCAmelCase_ ( self : Dict , _lowerCAmelCase : Union[str, Any] ): raise NotImplementedError('This model does not support the setting of `num_blocks`. Please set `block_sizes`.' )
31
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
0
import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __UpperCamelCase : def __init__( self , _UpperCamelCase , _UpperCamelCase=13 , _UpperCamelCase=10 , _UpperCamelCase=3 , _UpperCamelCase=2 , _UpperCamelCase=2 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=32 , _UpperCamelCase=5 , _UpperCamelCase=4 , _UpperCamelCase=37 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=10 , _UpperCamelCase=0.02 , _UpperCamelCase="divided_space_time" , _UpperCamelCase=None , ): _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = patch_size _UpperCAmelCase = num_frames _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = attention_type _UpperCAmelCase = initializer_range _UpperCAmelCase = scope _UpperCAmelCase = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token _UpperCAmelCase = (image_size // patch_size) ** 2 _UpperCAmelCase = (num_frames) * self.num_patches_per_frame + 1 def UpperCamelCase( self ): _UpperCAmelCase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase( self ): _UpperCAmelCase = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) _UpperCAmelCase = self.num_labels return config def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): _UpperCAmelCase = TimesformerModel(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() _UpperCAmelCase = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): _UpperCAmelCase = TimesformerForVideoClassification(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() _UpperCAmelCase = model(_UpperCamelCase ) # verify the logits shape _UpperCAmelCase = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , _UpperCamelCase ) def UpperCamelCase( self ): _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( A__ , A__ , unittest.TestCase ): __A : List[Any] = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () __A : Union[str, Any] = ( {"""feature-extraction""": TimesformerModel, """video-classification""": TimesformerForVideoClassification} if is_torch_available() else {} ) __A : int = False __A : List[str] = False __A : int = False __A : str = False def UpperCamelCase( self ): _UpperCAmelCase = TimesformerModelTester(self ) _UpperCAmelCase = ConfigTester( self , config_class=_UpperCamelCase , has_text_modality=_UpperCamelCase , hidden_size=37 ) def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=False ): _UpperCAmelCase = copy.deepcopy(_UpperCamelCase ) if return_labels: if model_class in get_values(_UpperCamelCase ): _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_UpperCamelCase ) return inputs_dict def UpperCamelCase( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''TimeSformer does not use inputs_embeds''' ) def UpperCamelCase( self ): pass def UpperCamelCase( self ): _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(_UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_UpperCamelCase , nn.Linear ) ) def UpperCamelCase( self ): _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(_UpperCamelCase ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _UpperCamelCase ) def UpperCamelCase( self ): _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def UpperCamelCase( self ): _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*_UpperCamelCase ) @slow def UpperCamelCase( self ): for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = TimesformerModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) def UpperCamelCase( self ): if not self.has_attentions: pass else: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = True for model_class in self.all_model_classes: _UpperCAmelCase = self.model_tester.seq_length _UpperCAmelCase = self.model_tester.num_frames _UpperCAmelCase = True _UpperCAmelCase = False _UpperCAmelCase = True _UpperCAmelCase = model_class(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _UpperCAmelCase = True _UpperCAmelCase = model_class(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) _UpperCAmelCase = len(_UpperCamelCase ) # Check attention is always last and order is fine _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = model_class(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(out_len + 1 , len(_UpperCamelCase ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def UpperCamelCase( self ): def check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): _UpperCAmelCase = model_class(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(_UpperCamelCase ) , _UpperCamelCase ) _UpperCAmelCase = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def A__ ( ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) _UpperCAmelCase = np.load(SCREAMING_SNAKE_CASE_ ) return list(SCREAMING_SNAKE_CASE_ ) @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def UpperCamelCase( self ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def UpperCamelCase( self ): _UpperCAmelCase = TimesformerForVideoClassification.from_pretrained('''facebook/timesformer-base-finetuned-k400''' ).to( _UpperCamelCase ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_video() _UpperCAmelCase = image_processor(video[:8] , return_tensors='''pt''' ).to(_UpperCamelCase ) # forward pass with torch.no_grad(): _UpperCAmelCase = model(**_UpperCamelCase ) # verify the logits _UpperCAmelCase = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape , _UpperCamelCase ) _UpperCAmelCase = torch.tensor([-0.3016, -0.7713, -0.4205] ).to(_UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _UpperCamelCase , atol=1e-4 ) )
32
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
0
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer lowerCamelCase__ : Tuple = """bart""" lowerCamelCase__ : Dict = True @st.cache(allow_output_mutation=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: if LOAD_DENSE_INDEX: snake_case__ = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) snake_case__ = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) snake_case__ = qar_model.eval() else: snake_case__ , snake_case__ = (None, None) if MODEL_TYPE == "bart": snake_case__ = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) snake_case__ = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) snake_case__ = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) snake_case__ = sas_model.eval() else: snake_case__ , snake_case__ = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> int: if LOAD_DENSE_INDEX: snake_case__ = faiss.StandardGpuResources() snake_case__ = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] snake_case__ = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 128) , ) snake_case__ = faiss.IndexFlatIP(128 ) snake_case__ = faiss.index_cpu_to_gpu(__lowerCAmelCase , 1 , __lowerCAmelCase ) wikiaab_gpu_index_flat.add(__lowerCAmelCase ) # TODO fix for larger GPU else: snake_case__ , snake_case__ = (None, None) snake_case__ = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> List[str]: snake_case__ = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) snake_case__ = elia['''train_eli5'''] snake_case__ = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 128) ) snake_case__ = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(__lowerCAmelCase ) return (elia_train, eli5_train_q_index) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = load_indexes() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[str] = load_models() lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = load_train_data() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase=10 ) -> List[Any]: snake_case__ = embed_questions_for_retrieval([question] , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ , snake_case__ = eli5_train_q_index.search(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ = [elia_train[int(__lowerCAmelCase )] for i in I[0]] return nn_examples def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase="wiki40b" , __lowerCAmelCase="dense" , __lowerCAmelCase=10 ) -> int: if source == "none": snake_case__ , snake_case__ = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": snake_case__ , snake_case__ = query_qa_dense_index( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) else: snake_case__ , snake_case__ = query_es_index( __lowerCAmelCase , __lowerCAmelCase , index_name='''english_wiki40b_snippets_100w''' , n_results=__lowerCAmelCase , ) snake_case__ = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] snake_case__ = '''question: {} context: {}'''.format(__lowerCAmelCase , __lowerCAmelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda __lowerCAmelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda __lowerCAmelCase : None), } ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=64 , __lowerCAmelCase=256 , __lowerCAmelCase=False , __lowerCAmelCase=2 , __lowerCAmelCase=0.95 , __lowerCAmelCase=0.8 ) -> Dict: with torch.no_grad(): snake_case__ = qa_sas_generate( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , num_answers=1 , num_beams=__lowerCAmelCase , min_len=__lowerCAmelCase , max_len=__lowerCAmelCase , do_sample=__lowerCAmelCase , temp=__lowerCAmelCase , top_p=__lowerCAmelCase , top_k=__lowerCAmelCase , max_input_length=1024 , device='''cuda:0''' , )[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar lowerCamelCase__ : int = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" lowerCamelCase__ : List[Any] = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia lowerCamelCase__ : Any = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) lowerCamelCase__ : str = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] lowerCamelCase__ : Tuple = st.sidebar.checkbox("""Demo options""") if demo_options: lowerCamelCase__ : List[str] = st.sidebar.selectbox( """""", action_list, index=3, ) lowerCamelCase__ : Union[str, Any] = action_list.index(action_st) lowerCamelCase__ : List[Any] = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) lowerCamelCase__ : int = show_type == """Show full text of passages""" else: lowerCamelCase__ : str = 3 lowerCamelCase__ : Any = True lowerCamelCase__ : Tuple = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: lowerCamelCase__ : Optional[int] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) lowerCamelCase__ : Optional[int] = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) lowerCamelCase__ : int = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: lowerCamelCase__ : str = """wiki40b""" lowerCamelCase__ : str = """dense""" lowerCamelCase__ : int = """beam""" lowerCamelCase__ : Optional[int] = 2 lowerCamelCase__ : int = 6_4 lowerCamelCase__ : List[Any] = 2_5_6 lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : List[str] = None lowerCamelCase__ : str = st.sidebar.checkbox("""Generation options""") if generate_options: lowerCamelCase__ : int = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) lowerCamelCase__ : List[str] = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) lowerCamelCase__ : Optional[Any] = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=2_5_6, value=6_4, step=8, format=None, key=None ) lowerCamelCase__ : Optional[int] = st.sidebar.slider( """Maximum generation length""", min_value=6_4, max_value=5_1_2, value=2_5_6, step=1_6, format=None, key=None ) if sampled == "beam": lowerCamelCase__ : Optional[int] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: lowerCamelCase__ : Optional[Any] = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.9_5, step=0.0_1, format=None, key=None ) lowerCamelCase__ : Union[str, Any] = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.0_1, format=None, key=None ) lowerCamelCase__ : Dict = None # start main text lowerCamelCase__ : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] lowerCamelCase__ : Optional[Any] = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": lowerCamelCase__ : Dict = st.text_input("""Enter your question here:""", """""") else: lowerCamelCase__ : Union[str, Any] = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": lowerCamelCase__ , lowerCamelCase__ : List[str] = make_support(question, source=wiki_source, method="""dense""", n_results=1_0) lowerCamelCase__ , lowerCamelCase__ : Any = make_support(question, source=wiki_source, method="""sparse""", n_results=1_0) lowerCamelCase__ : Optional[Any] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] lowerCamelCase__ : Union[str, Any] = support_list[:1_0] lowerCamelCase__ : Optional[Any] = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = make_support(question, source=wiki_source, method=index_type, n_results=1_0) if action in [0, 3]: lowerCamelCase__ , lowerCamelCase__ : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): lowerCamelCase__ : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) lowerCamelCase__ : List[Any] = res[1].strip() if sec_titles == "": lowerCamelCase__ : int = """[{}]({})""".format(res[0], wiki_url) else: lowerCamelCase__ : str = sec_titles.split(""" & """) lowerCamelCase__ : int = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: lowerCamelCase__ : List[Any] = find_nearest_training(question) lowerCamelCase__ : int = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) lowerCamelCase__ : Optional[int] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) lowerCamelCase__ : Union[str, Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
33
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( a_ ): _A : Optional[int] = 'facebook/bart-large-mnli' _A : Union[str, Any] = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) _A : Dict = 'text_classifier' _A : Union[str, Any] = AutoTokenizer _A : Tuple = AutoModelForSequenceClassification _A : Optional[int] = ['text', ['text']] _A : Dict = ['text'] def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setup() UpperCAmelCase = self.model.config UpperCAmelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = labels return self.pre_processor( [text] * len(snake_case__ ) , [f'''This example is {label}''' for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def UpperCamelCase_ ( self , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = outputs.logits UpperCAmelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
673
0
"""simple docstring""" import random class snake_case_ : """simple docstring""" @staticmethod def UpperCAmelCase__ ( lowerCamelCase_) -> tuple[list[int], list[int]]: UpperCamelCase = [ord(lowerCamelCase_) for i in text] UpperCamelCase = [] UpperCamelCase = [] for i in plain: UpperCamelCase = random.randint(1 , 3_0_0) UpperCamelCase = (i + k) * k cipher.append(lowerCamelCase_) key.append(lowerCamelCase_) return cipher, key @staticmethod def UpperCAmelCase__ ( lowerCamelCase_ , lowerCamelCase_) -> str: UpperCamelCase = [] for i in range(len(lowerCamelCase_)): UpperCamelCase = int((cipher[i] - (key[i]) ** 2) / key[i]) plain.append(chr(lowerCamelCase_)) return "".join(lowerCamelCase_) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
34
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase_ ( a_ ): _A : Union[List[PIL.Image.Image], np.ndarray] _A : Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
673
0
from __future__ import annotations def a ( A__ ) -> bool: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = str(A__ ) return len(A__ ) == 9 and set(A__ ) == set('''123456789''' ) def a ( ) -> int | None: '''simple docstring''' for base_num in range(9_9_9_9 , 4_9_9_9 , -1 ): SCREAMING_SNAKE_CASE__ : Optional[Any] = 1_0_0_0_0_2 * base_num if is_9_pandigital(A__ ): return candidate for base_num in range(3_3_3 , 9_9 , -1 ): SCREAMING_SNAKE_CASE__ : Tuple = 1_0_0_2_0_0_3 * base_num if is_9_pandigital(A__ ): return candidate return None if __name__ == "__main__": print(F'''{solution() = }''')
35
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : Any = { '''configuration_encodec''': [ '''ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EncodecConfig''', ], '''feature_extraction_encodec''': ['''EncodecFeatureExtractor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ '''ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EncodecModel''', '''EncodecPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys lowerCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
0
from __future__ import annotations import math class _A : '''simple docstring''' def __init__( self ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' snake_case : Dict = size # approximate the overall size of segment tree with given value snake_case : Dict = [0 for i in range(0 ,4 * size )] # create array to store lazy update snake_case : List[Any] = [0 for i in range(0 ,4 * size )] snake_case : Any = [0 for i in range(0 ,4 * size )] # flag for lazy update def snake_case_ ( self ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return idx * 2 def snake_case_ ( self ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return idx * 2 + 1 def snake_case_ ( self ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if left_element == right_element: snake_case : int = a[left_element - 1] else: snake_case : List[str] = (left_element + right_element) // 2 self.build(self.left(SCREAMING_SNAKE_CASE_ ) ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) self.build(self.right(SCREAMING_SNAKE_CASE_ ) ,mid + 1 ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) snake_case : str = max( self.segment_tree[self.left(SCREAMING_SNAKE_CASE_ )] ,self.segment_tree[self.right(SCREAMING_SNAKE_CASE_ )] ) def snake_case_ ( self ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if self.flag[idx] is True: snake_case : int = self.lazy[idx] snake_case : List[str] = False if left_element != right_element: snake_case : int = self.lazy[idx] snake_case : List[str] = self.lazy[idx] snake_case : List[Any] = True snake_case : Union[str, Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: snake_case : Optional[Any] = val if left_element != right_element: snake_case : str = val snake_case : Optional[Any] = val snake_case : Optional[Any] = True snake_case : List[Any] = True return True snake_case : List[str] = (left_element + right_element) // 2 self.update(self.left(SCREAMING_SNAKE_CASE_ ) ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) self.update(self.right(SCREAMING_SNAKE_CASE_ ) ,mid + 1 ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) snake_case : Tuple = max( self.segment_tree[self.left(SCREAMING_SNAKE_CASE_ )] ,self.segment_tree[self.right(SCREAMING_SNAKE_CASE_ )] ) return True def snake_case_ ( self ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if self.flag[idx] is True: snake_case : List[Any] = self.lazy[idx] snake_case : List[Any] = False if left_element != right_element: snake_case : List[str] = self.lazy[idx] snake_case : int = self.lazy[idx] snake_case : int = True snake_case : str = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] snake_case : List[Any] = (left_element + right_element) // 2 snake_case : List[str] = self.query(self.left(SCREAMING_SNAKE_CASE_ ) ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) snake_case : Union[str, Any] = self.query(self.right(SCREAMING_SNAKE_CASE_ ) ,mid + 1 ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) return max(SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) def __str__( self ): '''simple docstring''' return str([self.query(1 ,1 ,self.size ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ) for i in range(1 ,self.size + 1 )] ) if __name__ == "__main__": __lowercase : Union[str, Any] = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] __lowercase : Union[str, Any] = 15 __lowercase : Union[str, Any] = SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
36
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
673
0
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class A__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[Any]=7 , lowerCamelCase__ : Any=3 , lowerCamelCase__ : Dict=18 , lowerCamelCase__ : str=30 , lowerCamelCase__ : Tuple=400 , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[Any]=None , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=False , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Optional[Any]=[0.5, 0.5, 0.5] , lowerCamelCase__ : Optional[Any]=[0.5, 0.5, 0.5] , ): a__ : Any = parent a__ : Any = batch_size a__ : Any = num_channels a__ : Union[str, Any] = image_size a__ : Optional[Any] = min_resolution a__ : str = max_resolution a__ : List[str] = do_resize a__ : Any = size if size is not None else {"height": 18, "width": 20} a__ : Union[str, Any] = do_thumbnail a__ : Any = do_align_axis a__ : List[Any] = do_pad a__ : List[str] = do_normalize a__ : Any = image_mean a__ : List[Any] = image_std def _UpperCamelCase( self : Union[str, Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class A__ ( A__ , unittest.TestCase ): """simple docstring""" _lowercase = DonutImageProcessor if is_vision_available() else None def _UpperCamelCase( self : Tuple ): a__ : int = DonutImageProcessingTester(self ) @property def _UpperCamelCase( self : List[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase( self : List[Any] ): a__ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "size" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "do_thumbnail" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "do_align_long_axis" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "do_pad" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "image_std" ) ) def _UpperCamelCase( self : Dict ): a__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 20} ) a__ : Any = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) # Previous config had dimensions in (width, height) order a__ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"height": 84, "width": 42} ) def _UpperCamelCase( self : Any ): pass @is_flaky() def _UpperCamelCase( self : List[Any] ): # Initialize image_processing a__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a__ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input a__ : List[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched a__ : Tuple = image_processing(lowerCamelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def _UpperCamelCase( self : Optional[int] ): # Initialize image_processing a__ : str = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors a__ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) # Test not batched input a__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched a__ : Optional[Any] = image_processing(lowerCamelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def _UpperCamelCase( self : str ): # Initialize image_processing a__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors a__ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input a__ : List[str] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched a__ : str = image_processing(lowerCamelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
37
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = VideoToVideoSDPipeline _A : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} _A : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} _A : int = PipelineTesterMixin.required_optional_params - {'latents'} _A : List[str] = False # No `output_type`. _A : Any = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) UpperCAmelCase = CLIPTextModel(snake_case__ ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase_ ( self , snake_case__ , snake_case__=0 ) -> List[str]: """simple docstring""" UpperCAmelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(snake_case__ ) else: UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = VideoToVideoSDPipeline(**snake_case__ ) UpperCAmelCase = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) UpperCAmelCase = """np""" UpperCAmelCase = sd_pipe(**snake_case__ ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case__ ) UpperCAmelCase = video.to("""cuda""" ) UpperCAmelCase = """Spiderman is surfing""" UpperCAmelCase = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type="""pt""" ).frames UpperCAmelCase = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
673
0
'''simple docstring''' import math def UpperCamelCase__ ( __magic_name__ : int ) -> int: '''simple docstring''' if not isinstance(__magic_name__ , __magic_name__ ): snake_case__ : Union[str, Any] = f"Input value of [number={number}] must be an integer" raise TypeError(__magic_name__ ) if number < 1: snake_case__ : Optional[int] = f"Input value of [number={number}] must be > 0" raise ValueError(__magic_name__ ) elif number == 1: return 3 elif number == 2: return 5 else: snake_case__ : Optional[int] = int(math.log(number // 3 , 2 ) ) + 2 snake_case__ : str = [3, 5] snake_case__ : List[Any] = 2 snake_case__ : Optional[Any] = 3 for block in range(1 , __magic_name__ ): for _ in range(__magic_name__ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): A_ : Optional[int] = 0 try: A_ : Dict = proth(number) except ValueError: print(F'ValueError: there is no {number}th Proth number') continue print(F'The {number}th Proth number: {value}')
38
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class UpperCamelCase_ ( a_ ): _A : int = 'wav2vec2' def __init__( self , snake_case__=32 , snake_case__=7_68 , snake_case__=12 , snake_case__=12 , snake_case__=30_72 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=1e-5 , snake_case__="group" , snake_case__="gelu" , snake_case__=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case__=(5, 2, 2, 2, 2, 2, 2) , snake_case__=(10, 3, 3, 3, 3, 2, 2) , snake_case__=False , snake_case__=1_28 , snake_case__=16 , snake_case__=False , snake_case__=True , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__=3_20 , snake_case__=2 , snake_case__=0.1 , snake_case__=1_00 , snake_case__=2_56 , snake_case__=2_56 , snake_case__=0.1 , snake_case__="sum" , snake_case__=False , snake_case__=False , snake_case__=2_56 , snake_case__=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case__=(5, 3, 3, 1, 1) , snake_case__=(1, 2, 3, 1, 1) , snake_case__=5_12 , snake_case__=0 , snake_case__=1 , snake_case__=2 , snake_case__=False , snake_case__=3 , snake_case__=2 , snake_case__=3 , snake_case__=None , snake_case__=None , **snake_case__ , ) -> Optional[Any]: """simple docstring""" super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) UpperCAmelCase = hidden_size UpperCAmelCase = feat_extract_norm UpperCAmelCase = feat_extract_activation UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = conv_bias UpperCAmelCase = num_conv_pos_embeddings UpperCAmelCase = num_conv_pos_embedding_groups UpperCAmelCase = len(self.conv_dim ) UpperCAmelCase = num_hidden_layers UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = feat_proj_dropout UpperCAmelCase = final_dropout UpperCAmelCase = layerdrop UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = vocab_size UpperCAmelCase = do_stable_layer_norm UpperCAmelCase = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase = apply_spec_augment UpperCAmelCase = mask_time_prob UpperCAmelCase = mask_time_length UpperCAmelCase = mask_time_min_masks UpperCAmelCase = mask_feature_prob UpperCAmelCase = mask_feature_length UpperCAmelCase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCAmelCase = num_codevectors_per_group UpperCAmelCase = num_codevector_groups UpperCAmelCase = contrastive_logits_temperature UpperCAmelCase = feat_quantizer_dropout UpperCAmelCase = num_negatives UpperCAmelCase = codevector_dim UpperCAmelCase = proj_codevector_dim UpperCAmelCase = diversity_loss_weight # ctc loss UpperCAmelCase = ctc_loss_reduction UpperCAmelCase = ctc_zero_infinity # adapter UpperCAmelCase = add_adapter UpperCAmelCase = adapter_kernel_size UpperCAmelCase = adapter_stride UpperCAmelCase = num_adapter_layers UpperCAmelCase = output_hidden_size or hidden_size UpperCAmelCase = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = list(snake_case__ ) UpperCAmelCase = xvector_output_dim @property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
673
0
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch lowerCAmelCase_ = '''sshleifer/bart-tiny-random''' lowerCAmelCase_ = '''patrickvonplaten/t5-tiny-random''' @require_torch class snake_case_ ( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__( self : List[str] ) ->Union[str, Any]: return AutoConfig.from_pretrained(_UpperCamelCase ) def snake_case__( self : List[str] ) ->str: snake_case_, *snake_case_ = create_student_by_copying_alternating_layers(_UpperCamelCase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def snake_case__( self : List[Any] ) ->Any: snake_case_, *snake_case_ = create_student_by_copying_alternating_layers(_UpperCamelCase , tempfile.mkdtemp() , e=1 , d=_UpperCamelCase ) def snake_case__( self : List[str] ) ->str: snake_case_, *snake_case_ = create_student_by_copying_alternating_layers(_UpperCamelCase , tempfile.mkdtemp() , e=1 , d=_UpperCamelCase ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def snake_case__( self : List[Any] ) ->int: snake_case_, *snake_case_ = create_student_by_copying_alternating_layers(_UpperCamelCase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def snake_case__( self : Tuple ) ->Any: with self.assertRaises(_UpperCamelCase ): create_student_by_copying_alternating_layers(_UpperCamelCase , tempfile.mkdtemp() , e=_UpperCamelCase , d=_UpperCamelCase )
39
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCAmelCase_ : Optional[Any] = NewType('''DataClass''', Any) lowerCAmelCase_ : Any = NewType('''DataClassType''', Any) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = {str(lowerCAmelCase ): choice for choice in choices} return lambda lowerCAmelCase : str_to_choice.get(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( *, lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = dataclasses.MISSING , lowerCAmelCase = None , **lowerCAmelCase , ): '''simple docstring''' if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCAmelCase = {} if aliases is not None: UpperCAmelCase = aliases if help is not None: UpperCAmelCase = help return dataclasses.field(metadata=lowerCAmelCase , default=lowerCAmelCase , default_factory=lowerCAmelCase , **lowerCAmelCase ) class UpperCamelCase_ ( a_ ): _A : Iterable[DataClassType] def __init__( self , snake_case__ , **snake_case__ ) -> List[str]: """simple docstring""" if "formatter_class" not in kwargs: UpperCAmelCase = ArgumentDefaultsHelpFormatter super().__init__(**snake_case__ ) if dataclasses.is_dataclass(snake_case__ ): UpperCAmelCase = [dataclass_types] UpperCAmelCase = list(snake_case__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(snake_case__ ) @staticmethod def UpperCamelCase_ ( snake_case__ , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = f'''--{field.name}''' UpperCAmelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , snake_case__ ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) UpperCAmelCase = kwargs.pop("""aliases""" , [] ) if isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [aliases] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(snake_case__ , """UnionType""" ) and isinstance(snake_case__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(snake_case__ ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" f''' Problem encountered in field \'{field.name}\'.''' ) if type(snake_case__ ) not in field.type.__args__: # filter `str` in Union UpperCAmelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCAmelCase = ( field.type.__args__[0] if isinstance(snake_case__ , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCAmelCase = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCAmelCase = {} if origin_type is Literal or (isinstance(field.type , snake_case__ ) and issubclass(field.type , snake_case__ )): if origin_type is Literal: UpperCAmelCase = field.type.__args__ else: UpperCAmelCase = [x.value for x in field.type] UpperCAmelCase = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default else: UpperCAmelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCAmelCase = copy(snake_case__ ) # Hack because type=bool in argparse does not behave as we want. UpperCAmelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCAmelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCAmelCase = default # This tells argparse we accept 0 or 1 value after --field_name UpperCAmelCase = """?""" # This is the value that will get picked if we do --field_name (without value) UpperCAmelCase = True elif isclass(snake_case__ ) and issubclass(snake_case__ , snake_case__ ): UpperCAmelCase = field.type.__args__[0] UpperCAmelCase = """+""" if field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() elif field.default is dataclasses.MISSING: UpperCAmelCase = True else: UpperCAmelCase = field.type if field.default is not dataclasses.MISSING: UpperCAmelCase = field.default elif field.default_factory is not dataclasses.MISSING: UpperCAmelCase = field.default_factory() else: UpperCAmelCase = True parser.add_argument(snake_case__ , *snake_case__ , **snake_case__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCAmelCase = False parser.add_argument(f'''--no_{field.name}''' , action="""store_false""" , dest=field.name , **snake_case__ ) def UpperCamelCase_ ( self , snake_case__ ) -> Any: """simple docstring""" if hasattr(snake_case__ , """_argument_group_name""" ): UpperCAmelCase = self.add_argument_group(dtype._argument_group_name ) else: UpperCAmelCase = self try: UpperCAmelCase = get_type_hints(snake_case__ ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(snake_case__ ): UpperCAmelCase = """.""".join(map(snake_case__ , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(snake_case__ ): if not field.init: continue UpperCAmelCase = type_hints[field.name] self._parse_dataclass_field(snake_case__ , snake_case__ ) def UpperCamelCase_ ( self , snake_case__=None , snake_case__=False , snake_case__=True , snake_case__=None , snake_case__=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCAmelCase = [] if args_filename: args_files.append(Path(snake_case__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCAmelCase = ArgumentParser() args_file_parser.add_argument(snake_case__ , type=snake_case__ , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCAmelCase , UpperCAmelCase = args_file_parser.parse_known_args(args=snake_case__ ) UpperCAmelCase = vars(snake_case__ ).get(args_file_flag.lstrip("""-""" ) , snake_case__ ) if cmd_args_file_paths: args_files.extend([Path(snake_case__ ) for p in cmd_args_file_paths] ) UpperCAmelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCAmelCase = file_args + args if args is not None else file_args + sys.argv[1:] UpperCAmelCase , UpperCAmelCase = self.parse_known_args(args=snake_case__ ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in vars(snake_case__ ).items() if k in keys} for k in keys: delattr(snake_case__ , snake_case__ ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(snake_case__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = set(args.keys() ) UpperCAmelCase = [] for dtype in self.dataclass_types: UpperCAmelCase = {f.name for f in dataclasses.fields(snake_case__ ) if f.init} UpperCAmelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCAmelCase = dtype(**snake_case__ ) outputs.append(snake_case__ ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(snake_case__ )}''' ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(snake_case__ ) , encoding="""utf-8""" ) as open_json_file: UpperCAmelCase = json.loads(open_json_file.read() ) UpperCAmelCase = self.parse_dict(snake_case__ , allow_extra_keys=snake_case__ ) return tuple(snake_case__ ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ = False ) -> Tuple[DataClass, ...]: """simple docstring""" UpperCAmelCase = self.parse_dict(yaml.safe_load(Path(snake_case__ ).read_text() ) , allow_extra_keys=snake_case__ ) return tuple(snake_case__ )
673
0
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer __UpperCAmelCase = ['''bert-base-uncased''', '''bert-base-cased'''] __UpperCAmelCase = '''hf-internal-testing/tiny-bert-tf-only''' if is_tf_available(): class lowerCAmelCase_ ( tf.keras.Model ): def __init__( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: super().__init__() UpperCamelCase : Optional[int] = tokenizer UpperCamelCase : List[str] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = TFAutoModel.from_config(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[Any]: UpperCamelCase : Optional[Any] = self.tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = self.bert(**SCREAMING_SNAKE_CASE_ ) return out["pooler_output"] @require_tf @require_tensorflow_text class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self ) -> Dict: super().setUp() UpperCamelCase : Optional[Any] = [ BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false UpperCamelCase : Optional[Any] = [TFBertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_, use_fast_bert_tokenizer=SCREAMING_SNAKE_CASE_ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCamelCase : str = [ 'This is a straightforward English test sentence.', 'This one has some weird characters\rto\nsee\r\nif those\u00E9break things.', 'Now we\'re going to add some Chinese: 一 二 三 一二三', 'And some much more rare Chinese: 齉 堃 齉堃', 'Je vais aussi écrire en français pour tester les accents', 'Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ', ] UpperCamelCase : Optional[int] = list(zip(self.test_sentences, self.test_sentences[::-1] ) ) def snake_case_ ( self ) -> Optional[Any]: for tokenizer, tf_tokenizer in zip(self.tokenizers, self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): UpperCamelCase : Optional[Any] = tokenizer(SCREAMING_SNAKE_CASE_, return_tensors='tf', padding='longest' ) UpperCamelCase : int = tf_tokenizer(SCREAMING_SNAKE_CASE_ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key], tf.intaa ) == tf_outputs[key] ) ) @slow def snake_case_ ( self ) -> int: for tf_tokenizer in self.tf_tokenizers: UpperCamelCase : Tuple = tf_tokenizer(self.paired_sentences ) UpperCamelCase : Union[str, Any] = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences], text_pair=[sentence[1] for sentence in self.paired_sentences], ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key], tf.intaa ) == separated_outputs[key] ) ) @slow def snake_case_ ( self ) -> Union[str, Any]: for tf_tokenizer in self.tf_tokenizers: UpperCamelCase : Optional[Any] = tf.function(SCREAMING_SNAKE_CASE_ ) for test_inputs in (self.test_sentences, self.paired_sentences): UpperCamelCase : Dict = tf.constant(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Union[str, Any] = compiled_tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Union[str, Any] = tf_tokenizer(SCREAMING_SNAKE_CASE_ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def snake_case_ ( self ) -> Tuple: for tf_tokenizer in self.tf_tokenizers: UpperCamelCase : Any = ModelToSave(tokenizer=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Any = tf.convert_to_tensor(self.test_sentences ) UpperCamelCase : Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCamelCase : Dict = Path(SCREAMING_SNAKE_CASE_ ) / 'saved.model' model.save(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[Any] = tf.keras.models.load_model(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[int] = loaded_model(SCREAMING_SNAKE_CASE_ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ), 1e-5 )
40
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCAmelCase_ : List[str] = False class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self , snake_case__=32 ) -> Optional[Any]: """simple docstring""" set_seed(0 ) UpperCAmelCase = UNetaDModel(sample_size=snake_case__ , in_channels=3 , out_channels=3 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) UpperCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=snake_case__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randn((4, 3, 32, 32) ).to(snake_case__ ) for _ in range(4 )] UpperCAmelCase = [torch.randint(0 , 10_00 , (4,) ).long().to(snake_case__ ) for _ in range(4 )] # train with a DDPM scheduler UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM UpperCAmelCase , UpperCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(snake_case__ ) for i in range(4 ): optimizer.zero_grad() UpperCAmelCase = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) UpperCAmelCase = model(snake_case__ , timesteps[i] ).sample UpperCAmelCase = torch.nn.functional.mse_loss(snake_case__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) )
673
0
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = (DDPMScheduler,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,**lowercase__ : int ): __lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', '''variance_type''': '''fixed_small''', '''clip_sample''': True, } config.update(**lowercase__ ) return config def SCREAMING_SNAKE_CASE ( self : Any ): for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] ,[0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=lowercase__ ,beta_end=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): self.check_over_configs(thresholding=lowercase__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=lowercase__ ,prediction_type=lowercase__ ,sample_max_value=lowercase__ ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for t in [0, 5_0_0, 9_9_9]: self.check_over_forward(time_step=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0_0_9_7_9 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.0_2 ) ) < 1e-5 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) __lowercase = len(lowercase__ ) __lowercase = self.dummy_model() __lowercase = self.dummy_sample_deter __lowercase = torch.manual_seed(0 ) for t in reversed(range(lowercase__ ) ): # 1. predict noise residual __lowercase = model(lowercase__ ,lowercase__ ) # 2. predict previous mean of sample x_t-1 __lowercase = scheduler.step(lowercase__ ,lowercase__ ,lowercase__ ,generator=lowercase__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance __lowercase = pred_prev_sample __lowercase = torch.sum(torch.abs(lowercase__ ) ) __lowercase = torch.mean(torch.abs(lowercase__ ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1e-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config(prediction_type='''v_prediction''' ) __lowercase = scheduler_class(**lowercase__ ) __lowercase = len(lowercase__ ) __lowercase = self.dummy_model() __lowercase = self.dummy_sample_deter __lowercase = torch.manual_seed(0 ) for t in reversed(range(lowercase__ ) ): # 1. predict noise residual __lowercase = model(lowercase__ ,lowercase__ ) # 2. predict previous mean of sample x_t-1 __lowercase = scheduler.step(lowercase__ ,lowercase__ ,lowercase__ ,generator=lowercase__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance __lowercase = pred_prev_sample __lowercase = torch.sum(torch.abs(lowercase__ ) ) __lowercase = torch.mean(torch.abs(lowercase__ ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1e-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) __lowercase = [1_0_0, 8_7, 5_0, 1, 0] scheduler.set_timesteps(timesteps=lowercase__ ) __lowercase = scheduler.timesteps for i, timestep in enumerate(lowercase__ ): if i == len(lowercase__ ) - 1: __lowercase = -1 else: __lowercase = timesteps[i + 1] __lowercase = scheduler.previous_timestep(lowercase__ ) __lowercase = prev_t.item() self.assertEqual(lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) __lowercase = [1_0_0, 8_7, 5_0, 5_1, 0] with self.assertRaises(lowercase__ ,msg='''`custom_timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) __lowercase = [1_0_0, 8_7, 5_0, 1, 0] __lowercase = len(lowercase__ ) with self.assertRaises(lowercase__ ,msg='''Can only pass one of `num_inference_steps` or `custom_timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=lowercase__ ,timesteps=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**lowercase__ ) __lowercase = [scheduler.config.num_train_timesteps] with self.assertRaises( lowercase__ ,msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' ,): scheduler.set_timesteps(timesteps=lowercase__ )
41
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class UpperCamelCase_ : def __init__( self , snake_case__=2 , snake_case__=3 , snake_case__=64 , snake_case__=None ) -> List[str]: """simple docstring""" UpperCAmelCase = np.random.default_rng(snake_case__ ) UpperCAmelCase = length UpperCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) UpperCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> int: """simple docstring""" return self.length def __getitem__( self , snake_case__ ) -> Tuple: """simple docstring""" return {"x": self.x[i], "y": self.y[i]} class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[str]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> List[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a[0] + self.b[0] class UpperCamelCase_ ( torch.nn.Module ): def __init__( self , snake_case__=0 , snake_case__=0 , snake_case__=False ) -> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = torch.nn.Parameter(torch.tensor(snake_case__ ).float() ) UpperCAmelCase = True def UpperCamelCase_ ( self , snake_case__=None ) -> Optional[Any]: """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) UpperCAmelCase = False return x * self.a + self.b def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = 16 ): '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer UpperCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) UpperCAmelCase = {"""train""": """tests/test_samples/MRPC/train.csv""", """validation""": """tests/test_samples/MRPC/dev.csv"""} UpperCAmelCase = load_dataset("""csv""" , data_files=lowerCAmelCase ) UpperCAmelCase = datasets["""train"""].unique("""label""" ) UpperCAmelCase = {v: i for i, v in enumerate(lowerCAmelCase )} def tokenize_function(lowerCAmelCase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase = tokenizer( examples["""sentence1"""] , examples["""sentence2"""] , truncation=lowerCAmelCase , max_length=lowerCAmelCase , padding="""max_length""" ) if "label" in examples: UpperCAmelCase = [label_to_id[l] for l in examples["""label"""]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase = datasets.map( lowerCAmelCase , batched=lowerCAmelCase , remove_columns=["""sentence1""", """sentence2""", """label"""] , ) def collate_fn(lowerCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return tokenizer.pad(lowerCAmelCase , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. UpperCAmelCase = DataLoader(tokenized_datasets["""train"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=2 ) UpperCAmelCase = DataLoader(tokenized_datasets["""validation"""] , shuffle=lowerCAmelCase , collate_fn=lowerCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
673
0
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input A_ = "Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine" def _UpperCamelCase ( ) -> str: lowerCamelCase_ = _ask_options( 'In which compute environment are you running?' ,['This machine', 'AWS (Amazon SageMaker)'] ,_convert_compute_environment ,) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: lowerCamelCase_ = get_sagemaker_input() else: lowerCamelCase_ = get_cluster_input() return config def _UpperCamelCase ( __UpperCamelCase=None ) -> Union[str, Any]: if subparsers is not None: lowerCamelCase_ = subparsers.add_parser('config' ,description=__UpperCamelCase ) else: lowerCamelCase_ = argparse.ArgumentParser('Accelerate config command' ,description=__UpperCamelCase ) parser.add_argument( '--config_file' ,default=__UpperCamelCase ,help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) ,) if subparsers is not None: parser.set_defaults(func=__UpperCamelCase ) return parser def _UpperCamelCase ( __UpperCamelCase ) -> Tuple: lowerCamelCase_ = get_user_input() if args.config_file is not None: lowerCamelCase_ = args.config_file else: if not os.path.isdir(__UpperCamelCase ): os.makedirs(__UpperCamelCase ) lowerCamelCase_ = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(__UpperCamelCase ) else: config.to_yaml_file(__UpperCamelCase ) print(f'''accelerate configuration saved at {config_file}''' ) def _UpperCamelCase ( ) -> List[Any]: lowerCamelCase_ = config_command_parser() lowerCamelCase_ = parser.parse_args() config_command(__UpperCamelCase ) if __name__ == "__main__": main()
42
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = hidden_states.shape UpperCAmelCase = jax.image.resize( snake_case__ , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , snake_case__ ) -> Any: """simple docstring""" UpperCAmelCase = self.conv(snake_case__ ) return hidden_states class UpperCamelCase_ ( nn.Module ): _A : int _A : int = None _A : float = 0.0 _A : bool = None _A : jnp.dtype = jnp.floataa def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.in_channels if self.out_channels is None else self.out_channels UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = nn.Dense(snake_case__ , dtype=self.dtype ) UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCAmelCase = nn.Dropout(self.dropout_prob ) UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCAmelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut UpperCAmelCase = None if use_nin_shortcut: UpperCAmelCase = nn.Conv( snake_case__ , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self , snake_case__ , snake_case__ , snake_case__=True ) -> List[Any]: """simple docstring""" UpperCAmelCase = hidden_states UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) UpperCAmelCase = self.time_emb_proj(nn.swish(snake_case__ ) ) UpperCAmelCase = jnp.expand_dims(jnp.expand_dims(snake_case__ , 1 ) , 1 ) UpperCAmelCase = hidden_states + temb UpperCAmelCase = self.norma(snake_case__ ) UpperCAmelCase = nn.swish(snake_case__ ) UpperCAmelCase = self.dropout(snake_case__ , snake_case__ ) UpperCAmelCase = self.conva(snake_case__ ) if self.conv_shortcut is not None: UpperCAmelCase = self.conv_shortcut(snake_case__ ) return hidden_states + residual
673
0
def _a ( SCREAMING_SNAKE_CASE = 10_00 ): """simple docstring""" lowercase__ = 2**power lowercase__ = str(SCREAMING_SNAKE_CASE ) lowercase__ = list(SCREAMING_SNAKE_CASE ) lowercase__ = 0 for i in list_num: sum_of_num += int(SCREAMING_SNAKE_CASE ) return sum_of_num if __name__ == "__main__": lowerCAmelCase = int(input('Enter the power of 2: ').strip()) print('2 ^ ', power, ' = ', 2**power) lowerCAmelCase = solution(power) print('Sum of the digits is: ', result)
43
"""simple docstring""" from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=30 , snake_case__=2 , snake_case__=3 , snake_case__=True , snake_case__=True , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=10 , snake_case__=0.02 , snake_case__=3 , snake_case__=None , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 1 def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModel(config=snake_case__ ) UpperCAmelCase = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) UpperCAmelCase = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase = self.image_size // 2 UpperCAmelCase = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFViTForImageClassification(snake_case__ ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[int] = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _A : Optional[Any] = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) _A : Optional[int] = False _A : Any = False _A : List[str] = False def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = TFViTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer ) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(snake_case__ ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(snake_case__ ) def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCamelCase_ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=snake_case__ , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**snake_case__ ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case__ ) UpperCAmelCase = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , snake_case__ , atol=1e-4 )
673
0
'''simple docstring''' def A_ ( _lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase : int = len(_lowerCAmelCase ) _lowerCamelCase : Dict = sum(_lowerCAmelCase ) _lowerCamelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _lowerCamelCase : Optional[Any] = True for i in range(1 , s + 1 ): _lowerCamelCase : Optional[int] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _lowerCamelCase : Tuple = dp[i][j - 1] if arr[i - 1] <= j: _lowerCamelCase : Optional[Any] = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _lowerCamelCase : int = s - 2 * j break return diff
44
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=5_12 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ , token_type_ids=snake_case__ ) UpperCAmelCase = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = NystromformerForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = NystromformerForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = NystromformerForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) -> Dict: """simple docstring""" UpperCAmelCase = self.num_choices UpperCAmelCase = NystromformerForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( a_ , a_ , unittest.TestCase ): _A : Optional[Any] = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) _A : Optional[Any] = ( { 'feature-extraction': NystromformerModel, 'fill-mask': NystromformerForMaskedLM, 'question-answering': NystromformerForQuestionAnswering, 'text-classification': NystromformerForSequenceClassification, 'token-classification': NystromformerForTokenClassification, 'zero-shot': NystromformerForSequenceClassification, } if is_torch_available() else {} ) _A : int = False _A : Dict = False def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = NystromformerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case__ ) def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case__ ) def UpperCamelCase_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case__ ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case__ ) @slow def UpperCamelCase_ ( self ) -> int: """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = NystromformerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): UpperCAmelCase = model(snake_case__ )[0] UpperCAmelCase = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , snake_case__ ) UpperCAmelCase = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) ) @slow def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = """the [MASK] of Belgium is Brussels""" UpperCAmelCase = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) UpperCAmelCase = tokenizer(snake_case__ , return_tensors="""pt""" ) with torch.no_grad(): UpperCAmelCase = model(encoding.input_ids ).logits UpperCAmelCase = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(snake_case__ ) , """capital""" )
673
0
import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece_with_bytefallback.model") @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( lowercase , unittest.TestCase ): """simple docstring""" _snake_case : List[str] = GPTSwaTokenizer _snake_case : Dict = False _snake_case : str = True _snake_case : int = False def __a ( self :Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase__ :int = GPTSwaTokenizer(lowerCamelCase__ , eos_token="""<unk>""" , bos_token="""<unk>""" , pad_token="""<unk>""" ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self :Dict , lowerCamelCase__ :Tuple ): UpperCamelCase__ :List[Any] = """This is a test""" UpperCamelCase__ :Optional[Any] = """This is a test""" return input_text, output_text def __a ( self :Optional[Any] ): UpperCamelCase__ :int = """<s>""" UpperCamelCase__ :Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def __a ( self :Union[str, Any] ): UpperCamelCase__ :Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<unk>""" ) self.assertEqual(vocab_keys[1] , """<s>""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(lowerCamelCase__ ) , 20_00 ) def __a ( self :List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size , 20_00 ) def __a ( self :Any ): UpperCamelCase__ :Union[str, Any] = GPTSwaTokenizer(lowerCamelCase__ ) UpperCamelCase__ :Optional[int] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(lowerCamelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [4_65, 2_87, 2_65, 6_31, 8_42] ) UpperCamelCase__ :Dict = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) # fmt: off self.assertListEqual( lowerCamelCase__ , ["""▁I""", """▁was""", """▁bor""", """n""", """▁in""", """▁""", """<0x39>""", """2""", """0""", """0""", """0""", """,""", """▁and""", """▁this""", """▁is""", """▁f""", """al""", """s""", """<0xC3>""", """<0xA9>""", """."""] , ) # fmt: on UpperCamelCase__ :Tuple = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [2_62, 2_72, 15_25, 2_86, 2_71, 2_68, 60, 9_16, 6_33, 6_33, 6_33, 2_59, 2_66, 3_01, 2_87, 3_84, 3_67, 2_63, 1_98, 1_72, 2_60] , ) UpperCamelCase__ :Optional[Any] = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) # fmt: off self.assertListEqual( lowerCamelCase__ , ["""▁I""", """▁was""", """▁bor""", """n""", """▁in""", """▁""", """<0x39>""", """2""", """0""", """0""", """0""", """,""", """▁and""", """▁this""", """▁is""", """▁f""", """al""", """s""", """<0xC3>""", """<0xA9>""", """."""] ) # fmt: on def __a ( self :Optional[Any] ): UpperCamelCase__ :Dict = GPTSwaTokenizer(lowerCamelCase__ ) UpperCamelCase__ :List[str] = ["""This is a test""", """I was born in 92000, and this is falsé."""] UpperCamelCase__ :Union[str, Any] = [ [4_65, 2_87, 2_65, 6_31, 8_42], [2_62, 2_72, 15_25, 2_86, 2_71, 2_68, 60, 9_16, 6_33, 6_33, 6_33, 2_59, 2_66, 3_01, 2_87, 3_84, 3_67, 2_63, 1_98, 1_72, 2_60], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertListEqual(tokenizer.encode_fast(lowerCamelCase__ ) , lowerCamelCase__ ) # Test that decode_fast returns the input text for text, token_ids in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(tokenizer.decode_fast(lowerCamelCase__ ) , lowerCamelCase__ ) @slow def __a ( self :Optional[int] ): UpperCamelCase__ :Tuple = [ """<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')""", """Hey there, how are you doing this fine day?""", """This is a text with a trailing spaces followed by a dot .""", """Häj sväjs lillebrör! =)""", """Det är inget fel på Mr. Cool""", ] # fmt: off UpperCamelCase__ :str = {"""input_ids""": [[6_34_23, 5, 68_11, 1_49_54, 2_82, 8_16, 38_21, 6_34_66, 6_34_25, 6_34_62, 18, 6_39_78, 6_78, 3_01, 13_20, 6_34_23, 6_34_55, 6_34_58, 18, 6_39_82, 42_46, 39_40, 19_01, 4_77_89, 55_47, 1_89_94], [1_96_30, 11_00, 6_34_46, 13_42, 6_33, 5_44, 44_88, 5_93, 51_02, 24_16, 6_34_95, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [16_52, 4_28, 2_68, 19_36, 5_15, 2_68, 5_85_93, 2_24_13, 91_06, 5_46, 2_68, 3_32_13, 6_39_79, 6_98, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_51_30, 6_34_50, 9_24, 6_34_49, 22_49, 40_62, 15_58, 3_18, 6_35_04, 2_14_98, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_09, 3_77, 28_27, 25_59, 3_32, 65_75, 6_34_43, 2_68_01, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name="""AI-Sweden/gpt-sw3-126m""" , sequences=lowerCamelCase__ , )
45
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Optional[int] = False def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return TrainCommand(lowerCAmelCase ) class UpperCamelCase_ ( a_ ): @staticmethod def UpperCamelCase_ ( snake_case__ ) -> int: """simple docstring""" UpperCAmelCase = parser.add_parser("""train""" , help="""CLI tool to train a model on a task.""" ) train_parser.add_argument( """--train_data""" , type=snake_case__ , required=snake_case__ , help="""path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.""" , ) train_parser.add_argument( """--column_label""" , type=snake_case__ , default=0 , help="""Column of the dataset csv file with example labels.""" ) train_parser.add_argument( """--column_text""" , type=snake_case__ , default=1 , help="""Column of the dataset csv file with example texts.""" ) train_parser.add_argument( """--column_id""" , type=snake_case__ , default=2 , help="""Column of the dataset csv file with example ids.""" ) train_parser.add_argument( """--skip_first_row""" , action="""store_true""" , help="""Skip the first row of the csv file (headers).""" ) train_parser.add_argument("""--validation_data""" , type=snake_case__ , default="""""" , help="""path to validation dataset.""" ) train_parser.add_argument( """--validation_split""" , type=snake_case__ , default=0.1 , help="""if validation dataset is not provided, fraction of train dataset to use as validation dataset.""" , ) train_parser.add_argument("""--output""" , type=snake_case__ , default="""./""" , help="""path to saved the trained model.""" ) train_parser.add_argument( """--task""" , type=snake_case__ , default="""text_classification""" , help="""Task to train the model on.""" ) train_parser.add_argument( """--model""" , type=snake_case__ , default="""bert-base-uncased""" , help="""Model's name or path to stored model.""" ) train_parser.add_argument("""--train_batch_size""" , type=snake_case__ , default=32 , help="""Batch size for training.""" ) train_parser.add_argument("""--valid_batch_size""" , type=snake_case__ , default=64 , help="""Batch size for validation.""" ) train_parser.add_argument("""--learning_rate""" , type=snake_case__ , default=3e-5 , help="""Learning rate.""" ) train_parser.add_argument("""--adam_epsilon""" , type=snake_case__ , default=1e-08 , help="""Epsilon for Adam optimizer.""" ) train_parser.set_defaults(func=snake_case__ ) def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = logging.get_logger("""transformers-cli/training""" ) UpperCAmelCase = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=snake_case__ ) UpperCAmelCase = args.output UpperCAmelCase = args.column_label UpperCAmelCase = args.column_text UpperCAmelCase = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": UpperCAmelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) UpperCAmelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) UpperCAmelCase = args.validation_split UpperCAmelCase = args.train_batch_size UpperCAmelCase = args.valid_batch_size UpperCAmelCase = args.learning_rate UpperCAmelCase = args.adam_epsilon def UpperCamelCase_ ( self ) -> Any: """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" raise NotImplementedError def UpperCamelCase_ ( self ) -> str: """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
673
0
"""simple docstring""" from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCamelCase_( _lowerCamelCase ) -> Tuple: '''simple docstring''' def is_in_circle(_lowerCamelCase , _lowerCamelCase ) -> bool: _lowerCamelCase : int = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle _lowerCamelCase : Tuple = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(_lowerCamelCase ) ) # The ratio of the area for circle to square is pi/4. _lowerCamelCase : List[Any] = proportion * 4 print(F"""The estimated value of pi is {pi_estimate}""" ) print(F"""The numpy value of pi is {pi}""" ) print(F"""The total error is {abs(pi - pi_estimate )}""" ) def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0.0 , _lowerCamelCase = 1.0 , ) -> float: '''simple docstring''' return mean( function_to_integrate(uniform(_lowerCamelCase , _lowerCamelCase ) ) for _ in range(_lowerCamelCase ) ) * (max_value - min_value) def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase = 0.0 , _lowerCamelCase = 1.0 ) -> None: '''simple docstring''' def identity_function(_lowerCamelCase ) -> float: return x _lowerCamelCase : Union[str, Any] = area_under_curve_estimator( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) _lowerCamelCase : Tuple = (max_value * max_value - min_value * min_value) / 2 print("******************" ) print(F"""Estimating area under y=x where x varies from {min_value} to {max_value}""" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {expected_value}""" ) print(F"""Total error is {abs(estimated_value - expected_value )}""" ) print("******************" ) def lowerCamelCase_( _lowerCamelCase ) -> None: '''simple docstring''' def function_to_integrate(_lowerCamelCase ) -> float: return sqrt(4.0 - x * x ) _lowerCamelCase : int = area_under_curve_estimator( _lowerCamelCase , _lowerCamelCase , 0.0 , 2.0 ) print("******************" ) print("Estimating pi using area_under_curve_estimator" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {pi}""" ) print(F"""Total error is {abs(estimated_value - pi )}""" ) print("******************" ) if __name__ == "__main__": import doctest doctest.testmod()
46
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class UpperCamelCase_ : def __init__( self , snake_case__ , snake_case__=sys.maxsize ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """bilinear""" UpperCAmelCase = max_size UpperCAmelCase = short_edge_length def __call__( self , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = [] for img in imgs: UpperCAmelCase , UpperCAmelCase = img.shape[:2] # later: provide list and randomly choose index for resize UpperCAmelCase = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img UpperCAmelCase = size * 1.0 / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size if max(snake_case__ , snake_case__ ) > self.max_size: UpperCAmelCase = self.max_size * 1.0 / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase = int(neww + 0.5 ) UpperCAmelCase = int(newh + 0.5 ) if img.dtype == np.uinta: UpperCAmelCase = Image.fromarray(snake_case__ ) UpperCAmelCase = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) UpperCAmelCase = np.asarray(snake_case__ ) else: UpperCAmelCase = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw UpperCAmelCase = nn.functional.interpolate( snake_case__ , (newh, neww) , mode=self.interp_method , align_corners=snake_case__ ).squeeze(0 ) img_augs.append(snake_case__ ) return img_augs class UpperCamelCase_ : def __init__( self , snake_case__ ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) UpperCAmelCase = cfg.INPUT.FORMAT UpperCAmelCase = cfg.SIZE_DIVISIBILITY UpperCAmelCase = cfg.PAD_VALUE UpperCAmelCase = cfg.INPUT.MAX_SIZE_TEST UpperCAmelCase = cfg.MODEL.DEVICE UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) UpperCAmelCase = lambda snake_case__ : (x - self.pixel_mean) / self.pixel_std def UpperCamelCase_ ( self , snake_case__ ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = tuple(max(snake_case__ ) for s in zip(*[img.shape for img in images] ) ) UpperCAmelCase = [im.shape[-2:] for im in images] UpperCAmelCase = [ nn.functional.pad( snake_case__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(snake_case__ , snake_case__ ) ] return torch.stack(snake_case__ ), torch.tensor(snake_case__ ) def __call__( self , snake_case__ , snake_case__=False ) -> Optional[Any]: """simple docstring""" with torch.no_grad(): if not isinstance(snake_case__ , snake_case__ ): UpperCAmelCase = [images] if single_image: assert len(snake_case__ ) == 1 for i in range(len(snake_case__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(snake_case__ , images.pop(snake_case__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( snake_case__ , torch.as_tensor(img_tensorize(images.pop(snake_case__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge UpperCAmelCase = torch.tensor([im.shape[:2] for im in images] ) UpperCAmelCase = self.aug(snake_case__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic UpperCAmelCase = [self.normalizer(snake_case__ ) for x in images] # now pad them to do the following operations UpperCAmelCase , UpperCAmelCase = self.pad(snake_case__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad UpperCAmelCase = torch.true_divide(snake_case__ , snake_case__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' assert torch.isfinite(lowerCAmelCase ).all(), "Box tensor contains infinite or NaN!" UpperCAmelCase , UpperCAmelCase = box_size tensor[:, 0].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 1].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 2].clamp_(min=0 , max=lowerCAmelCase ) tensor[:, 3].clamp_(min=0 , max=lowerCAmelCase )
673
0
import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class _UpperCamelCase( __lowerCamelCase ): __SCREAMING_SNAKE_CASE : Any = '''MCTCTFeatureExtractor''' __SCREAMING_SNAKE_CASE : Tuple = '''AutoTokenizer''' def __init__( self : int , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' super().__init__(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __a : int = self.feature_extractor __a : Union[str, Any] = False def __call__( self : Tuple , *SCREAMING_SNAKE_CASE__ : List[Any] , **SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if "raw_speech" in kwargs: warnings.warn('Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.' ) __a : Optional[Any] = kwargs.pop('raw_speech' ) else: __a : Union[str, Any] = kwargs.pop('audio' , SCREAMING_SNAKE_CASE__ ) __a : Tuple = kwargs.pop('sampling_rate' , SCREAMING_SNAKE_CASE__ ) __a : Any = kwargs.pop('text' , SCREAMING_SNAKE_CASE__ ) if len(SCREAMING_SNAKE_CASE__ ) > 0: __a : List[str] = args[0] __a : Tuple = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if audio is not None: __a : Union[str, Any] = self.feature_extractor(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ , sampling_rate=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if text is not None: __a : Union[str, Any] = self.tokenizer(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if text is None: return inputs elif audio is None: return encodings else: __a : List[str] = encodings['input_ids'] return inputs def __lowerCAmelCase ( self : Optional[int] , *SCREAMING_SNAKE_CASE__ : Optional[Any] , **SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' return self.tokenizer.batch_decode(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def __lowerCAmelCase ( self : Optional[Any] , *SCREAMING_SNAKE_CASE__ : List[Any] , **SCREAMING_SNAKE_CASE__ : List[str] ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor.pad(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) __a : Dict = kwargs.pop('input_features' , SCREAMING_SNAKE_CASE__ ) __a : Union[str, Any] = kwargs.pop('labels' , SCREAMING_SNAKE_CASE__ ) if len(SCREAMING_SNAKE_CASE__ ) > 0: __a : Any = args[0] __a : Union[str, Any] = args[1:] if input_features is not None: __a : Tuple = self.feature_extractor.pad(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if labels is not None: __a : int = self.tokenizer.pad(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if labels is None: return input_features elif input_features is None: return labels else: __a : str = labels['input_ids'] return input_features def __lowerCAmelCase ( self : Dict , *SCREAMING_SNAKE_CASE__ : List[str] , **SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' return self.tokenizer.decode(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @contextmanager def __lowerCAmelCase ( self : int ): '''simple docstring''' warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your audio inputs, or in a separate call.' ) __a : Any = True __a : str = self.tokenizer yield __a : List[str] = self.feature_extractor __a : Any = False
47
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : List[str] = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase = """""" else: UpperCAmelCase = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = dct.pop(lowerCAmelCase ) UpperCAmelCase = val def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase = 1000 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = int(deit_name[-6:-4] ) UpperCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): UpperCAmelCase = 192 UpperCAmelCase = 768 UpperCAmelCase = 12 UpperCAmelCase = 3 elif deit_name[9:].startswith("""small""" ): UpperCAmelCase = 384 UpperCAmelCase = 1536 UpperCAmelCase = 12 UpperCAmelCase = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 # load original model from timm UpperCAmelCase = timm.create_model(lowerCAmelCase , pretrained=lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase = timm_model.state_dict() UpperCAmelCase = create_rename_keys(lowerCAmelCase , lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) read_in_q_k_v(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model UpperCAmelCase = DeiTForImageClassificationWithTeacher(lowerCAmelCase ).eval() model.load_state_dict(lowerCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase = DeiTImageProcessor(size=lowerCAmelCase , crop_size=config.image_size ) UpperCAmelCase = image_processor(images=prepare_img() , return_tensors="""pt""" ) UpperCAmelCase = encoding["""pixel_values"""] UpperCAmelCase = model(lowerCAmelCase ) UpperCAmelCase = timm_model(lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase_ : str = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
673
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class A : def __init__( self : Optional[Any] , __magic_name__ : str , ): """simple docstring""" lowerCAmelCase__ = parent lowerCAmelCase__ = 13 lowerCAmelCase__ = 7 lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = 99 lowerCAmelCase__ = 32 lowerCAmelCase__ = 2 lowerCAmelCase__ = 4 lowerCAmelCase__ = 37 lowerCAmelCase__ = "gelu" lowerCAmelCase__ = 0.1 lowerCAmelCase__ = 0.1 lowerCAmelCase__ = 512 lowerCAmelCase__ = 16 lowerCAmelCase__ = 2 lowerCAmelCase__ = 0.02 lowerCAmelCase__ = 3 lowerCAmelCase__ = 4 lowerCAmelCase__ = None def __SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ = None if self.use_input_mask: lowerCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __SCREAMING_SNAKE_CASE ( self : int , __magic_name__ : str , __magic_name__ : int , __magic_name__ : List[str] , __magic_name__ : Optional[Any] , __magic_name__ : str , __magic_name__ : int ): """simple docstring""" lowerCAmelCase__ = TFDistilBertModel(config=__magic_name__ ) lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCAmelCase__ = model(__magic_name__ ) lowerCAmelCase__ = [input_ids, input_mask] lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __SCREAMING_SNAKE_CASE ( self : Tuple , __magic_name__ : Optional[Any] , __magic_name__ : Tuple , __magic_name__ : Optional[int] , __magic_name__ : List[Any] , __magic_name__ : List[Any] , __magic_name__ : Any ): """simple docstring""" lowerCAmelCase__ = TFDistilBertForMaskedLM(config=__magic_name__ ) lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __SCREAMING_SNAKE_CASE ( self : Optional[int] , __magic_name__ : Optional[int] , __magic_name__ : str , __magic_name__ : Tuple , __magic_name__ : Dict , __magic_name__ : str , __magic_name__ : List[Any] ): """simple docstring""" lowerCAmelCase__ = TFDistilBertForQuestionAnswering(config=__magic_name__ ) lowerCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, } lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __SCREAMING_SNAKE_CASE ( self : str , __magic_name__ : List[str] , __magic_name__ : List[str] , __magic_name__ : str , __magic_name__ : str , __magic_name__ : Optional[int] , __magic_name__ : Union[str, Any] ): """simple docstring""" lowerCAmelCase__ = self.num_labels lowerCAmelCase__ = TFDistilBertForSequenceClassification(__magic_name__ ) lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __SCREAMING_SNAKE_CASE ( self : List[Any] , __magic_name__ : Any , __magic_name__ : Optional[int] , __magic_name__ : Optional[Any] , __magic_name__ : List[str] , __magic_name__ : Optional[Any] , __magic_name__ : Tuple ): """simple docstring""" lowerCAmelCase__ = self.num_choices lowerCAmelCase__ = TFDistilBertForMultipleChoice(__magic_name__ ) lowerCAmelCase__ = tf.tile(tf.expand_dims(__magic_name__ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase__ = tf.tile(tf.expand_dims(__magic_name__ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase__ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __magic_name__ : List[Any] , __magic_name__ : int , __magic_name__ : Dict , __magic_name__ : List[Any] , __magic_name__ : str , __magic_name__ : Dict ): """simple docstring""" lowerCAmelCase__ = self.num_labels lowerCAmelCase__ = TFDistilBertForTokenClassification(__magic_name__ ) lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCAmelCase__ = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" lowerCAmelCase__ = self.prepare_config_and_inputs() ((lowerCAmelCase__) ,(lowerCAmelCase__) ,(lowerCAmelCase__) ,(lowerCAmelCase__) ,(lowerCAmelCase__) ,(lowerCAmelCase__)) = config_and_inputs lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): snake_case__ :List[Any] = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) snake_case__ :Dict = ( { 'feature-extraction': TFDistilBertModel, 'fill-mask': TFDistilBertForMaskedLM, 'question-answering': TFDistilBertForQuestionAnswering, 'text-classification': TFDistilBertForSequenceClassification, 'token-classification': TFDistilBertForTokenClassification, 'zero-shot': TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) snake_case__ :List[str] = False snake_case__ :Tuple = False def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" lowerCAmelCase__ = TFDistilBertModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=__magic_name__ , dim=37 ) def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def __SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*__magic_name__ ) @slow def __SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): lowerCAmelCase__ = TFDistilBertModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) @require_tf class A ( unittest.TestCase ): @slow def __SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" lowerCAmelCase__ = TFDistilBertModel.from_pretrained("distilbert-base-uncased" ) lowerCAmelCase__ = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase__ = model(__magic_name__ )[0] lowerCAmelCase__ = [1, 6, 768] self.assertEqual(output.shape , __magic_name__ ) lowerCAmelCase__ = tf.constant( [ [ [0.1926_1885, -0.1373_2955, 0.411_9799], [0.2215_0156, -0.0742_2661, 0.3903_7204], [0.2275_6018, -0.089_6414, 0.370_1467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __magic_name__ , atol=1E-4 )
48
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCamelCase_ ( unittest.TestCase ): def __init__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = 32 , snake_case__ = True , snake_case__ = 1 / 2_55 , snake_case__ = True , snake_case__ = True , snake_case__ = [0.48_145_466, 0.4_578_275, 0.40_821_073] , snake_case__ = [0.26_862_954, 0.26_130_258, 0.27_577_711] , snake_case__ = True , snake_case__=7 , snake_case__=30 , snake_case__=4_00 , snake_case__=3 , ) -> List[str]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = do_resize UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_88} UpperCAmelCase = size_divisor UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_pad UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCamelCase_ ( self , snake_case__ , snake_case__=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = image_inputs[0] if isinstance(snake_case__ , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] UpperCAmelCase = size / min(snake_case__ , snake_case__ ) if h < w: UpperCAmelCase , UpperCAmelCase = size, scale * w else: UpperCAmelCase , UpperCAmelCase = scale * h, size UpperCAmelCase = int((13_33 / 8_00) * size ) if max(snake_case__ , snake_case__ ) > max_size: UpperCAmelCase = max_size / max(snake_case__ , snake_case__ ) UpperCAmelCase = newh * scale UpperCAmelCase = neww * scale UpperCAmelCase , UpperCAmelCase = int(newh + 0.5 ), int(neww + 0.5 ) UpperCAmelCase , UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] UpperCAmelCase = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = BridgeTowerImageProcessingTester(self ) @property def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case__ , """image_std""" ) ) self.assertTrue(hasattr(snake_case__ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case__ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case__ , """size""" ) ) self.assertTrue(hasattr(snake_case__ , """size_divisor""" ) ) def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase = image_processing(snake_case__ , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
673
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def lowercase__ ( snake_case_ :Tuple ): __UpperCAmelCase = args.pruning_method __UpperCAmelCase = args.threshold __UpperCAmelCase = args.model_name_or_path.rstrip('''/''' ) __UpperCAmelCase = args.target_model_path print(F'''Load fine-pruned model from {model_name_or_path}''' ) __UpperCAmelCase = torch.load(os.path.join(snake_case_ , '''pytorch_model.bin''' ) ) __UpperCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: __UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: __UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) elif "bias" in name: __UpperCAmelCase = tensor print(F'''Copied layer {name}''' ) else: if pruning_method == "magnitude": __UpperCAmelCase = MagnitudeBinarizer.apply(inputs=snake_case_ , threshold=snake_case_ ) __UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue __UpperCAmelCase = name[:-6] __UpperCAmelCase = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase = TopKBinarizer.apply(snake_case_ , snake_case_ ) __UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue __UpperCAmelCase = name[:-6] __UpperCAmelCase = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase = ThresholdBinarizer.apply(snake_case_ , snake_case_ , snake_case_ ) __UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue __UpperCAmelCase = name[:-6] __UpperCAmelCase = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase , __UpperCAmelCase = -0.1, 1.1 __UpperCAmelCase = torch.sigmoid(snake_case_ ) __UpperCAmelCase = s * (r - l) + l __UpperCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) __UpperCAmelCase = tensor * mask print(F'''Pruned layer {name}''' ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: __UpperCAmelCase = os.path.join( os.path.dirname(snake_case_ ) , F'''bertarized_{os.path.basename(snake_case_ )}''' ) if not os.path.isdir(snake_case_ ): shutil.copytree(snake_case_ , snake_case_ ) print(F'''\nCreated folder {target_model_path}''' ) torch.save(snake_case_ , os.path.join(snake_case_ , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": _lowercase : Tuple = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) _lowercase : str = parser.parse_args() main(args)
49
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : List[str] = XLMRobertaTokenizer _A : List[str] = XLMRobertaTokenizerFast _A : Optional[Any] = True _A : List[str] = True def UpperCamelCase_ ( self ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(snake_case__ ) , 10_02 ) def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def UpperCamelCase_ ( self ) -> str: """simple docstring""" UpperCAmelCase = XLMRobertaTokenizer(snake_case__ , keep_accents=snake_case__ ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(snake_case__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(snake_case__ ) self.assertListEqual( snake_case__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(snake_case__ ) self.assertListEqual( snake_case__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase = (self.rust_tokenizer_class, """hf-internal-testing/tiny-xlm-roberta""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) UpperCAmelCase = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=True UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it save with the same files self.assertSequenceEqual(snake_case__ , snake_case__ ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) # Save tokenizer rust, legacy_format=False UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = tokenizer_r.save_pretrained(snake_case__ , legacy_format=snake_case__ ) UpperCAmelCase = tokenizer_p.save_pretrained(snake_case__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase = tokenizer_r.from_pretrained(snake_case__ ) UpperCAmelCase = tokenizer_p.from_pretrained(snake_case__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(snake_case__ , snake_case__ ) ) shutil.rmtree(snake_case__ ) @cached_property def UpperCamelCase_ ( self ) -> int: """simple docstring""" return XLMRobertaTokenizer.from_pretrained("""xlm-roberta-base""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(snake_case__ , f.name ) UpperCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=snake_case__ ) UpperCAmelCase = pickle.dumps(snake_case__ ) pickle.loads(snake_case__ ) def UpperCamelCase_ ( self ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = """I was born in 92000, and this is falsé.""" UpperCAmelCase = tokenizer.tokenize(snake_case__ ) UpperCAmelCase = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(snake_case__ ) UpperCAmelCase = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) @slow def UpperCamelCase_ ( self ) -> Dict: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) UpperCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(snake_case__ , self.big_tokenizer.encode(snake_case__ ) ) @slow def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="""xlm-roberta-base""" , revision="""d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3""" , )
673
0
'''simple docstring''' from collections.abc import Generator def A__ ( ): lowerCamelCase__ , lowerCamelCase__ = 0, 1 while True: lowerCamelCase__ , lowerCamelCase__ = b, a + b yield b def A__ ( __lowerCAmelCase : int = 1000 ): lowerCamelCase__ = 1 lowerCamelCase__ = fibonacci_generator() while len(str(next(__lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
50
"""simple docstring""" import socket def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCAmelCase = socket.gethostname() UpperCAmelCase = 12312 sock.connect((host, port) ) sock.send(b"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: UpperCAmelCase = sock.recv(1024 ) if not data: break out_file.write(lowerCAmelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
673
0
'''simple docstring''' def __snake_case ( SCREAMING_SNAKE_CASE_ : str ) -> bool: """simple docstring""" UpperCAmelCase = [int(SCREAMING_SNAKE_CASE_ ) for i in ip_va_address.split('''.''' ) if i.isdigit()] return len(SCREAMING_SNAKE_CASE_ ) == 4 and all(0 <= int(SCREAMING_SNAKE_CASE_ ) <= 254 for octet in octets ) if __name__ == "__main__": a__ : List[str] = input().strip() a__ : List[Any] = 'valid' if is_ip_va_address_valid(ip) else 'invalid' print(F"""{ip} is a {valid_or_invalid} IP v4 address.""")
51
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return math.sqrt(lowerCAmelCase ) * math.sqrt(lowerCAmelCase ) == num def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = 0 UpperCAmelCase = n while left <= right: UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase = mid - 1 else: UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
673
0
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation A = logging.get_logger(__name__) A = {'''tokenizer_file''': '''tokenizer.json'''} A = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class __lowercase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = ['''input_ids''', '''attention_mask'''] __lowerCAmelCase = None def __init__( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase="<unk>" , _UpperCAmelCase="<s>" , _UpperCAmelCase="</s>" , _UpperCAmelCase="<pad>" , _UpperCAmelCase=False , _UpperCAmelCase=False , **_UpperCAmelCase , ): super().__init__( _UpperCAmelCase , _UpperCAmelCase , tokenizer_file=_UpperCAmelCase , unk_token=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , add_prefix_space=_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase , **_UpperCAmelCase , ) __a : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , _UpperCAmelCase ) != add_prefix_space: __a : Tuple = getattr(_UpperCAmelCase , pre_tok_state.pop('''type''' ) ) __a : Optional[int] = add_prefix_space __a : str = pre_tok_class(**_UpperCAmelCase ) __a : Optional[int] = add_prefix_space def _lowerCamelCase ( self , *_UpperCAmelCase , **_UpperCAmelCase ): __a : str = kwargs.get('''is_split_into_words''' , _UpperCAmelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''' ) return super()._batch_encode_plus(*_UpperCAmelCase , **_UpperCAmelCase ) def _lowerCamelCase ( self , *_UpperCAmelCase , **_UpperCAmelCase ): __a : int = kwargs.get('''is_split_into_words''' , _UpperCAmelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''' ) return super()._encode_plus(*_UpperCAmelCase , **_UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase = None ): __a : Optional[int] = self._tokenizer.model.save(_UpperCAmelCase , name=_UpperCAmelCase ) return tuple(_UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase ): __a : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) + [self.eos_token_id] ) if len(_UpperCAmelCase ) > self.model_max_length: __a : Optional[int] = input_ids[-self.model_max_length :] return input_ids
52
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _lowerCAmelCase ( *lowerCAmelCase ): '''simple docstring''' if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase = list(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): UpperCAmelCase = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _lowerCAmelCase ( lowerCAmelCase = None , lowerCAmelCase = 128 ): '''simple docstring''' if function is None: return functools.partial(lowerCAmelCase , starting_batch_size=lowerCAmelCase ) UpperCAmelCase = starting_batch_size def decorator(*lowerCAmelCase , **lowerCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() UpperCAmelCase = list(inspect.signature(lowerCAmelCase ).parameters.keys() ) # Guard against user error if len(lowerCAmelCase ) < (len(lowerCAmelCase ) + 1): UpperCAmelCase = """, """.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) except Exception as e: if should_reduce_batch_size(lowerCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
673
0
import requests _snake_case : Tuple = '' # <-- Put your OpenWeatherMap appid here! _snake_case : Dict = 'https://api.openweathermap.org/data/2.5/' def a_ ( lowerCAmelCase_ : str = "Chicago", lowerCAmelCase_ : str = APPID ): return requests.get(URL_BASE + 'weather', params=locals() ).json() def a_ ( lowerCAmelCase_ : str = "Kolkata, India", lowerCAmelCase_ : str = APPID ): return requests.get(URL_BASE + 'forecast', params=locals() ).json() def a_ ( lowerCAmelCase_ : float = 55.68, lowerCAmelCase_ : float = 12.57, lowerCAmelCase_ : str = APPID ): return requests.get(URL_BASE + 'onecall', params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: _snake_case : List[str] = input('Enter a location:').strip() if location: pprint(current_weather(location)) else: break
53
"""simple docstring""" import math def _lowerCAmelCase ( lowerCAmelCase = 100 ): '''simple docstring''' UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
673
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def a__ ( lowercase__ ): '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class A ( __lowercase ): @staticmethod def lowerCAmelCase__ ( _lowerCAmelCase: ArgumentParser ) -> Tuple: '''simple docstring''' UpperCAmelCase_ =parser.add_parser("download" ) download_parser.add_argument( "--cache-dir" , type=_lowerCAmelCase , default=_lowerCAmelCase , help="Path to location to store the models" ) download_parser.add_argument( "--force" , action="store_true" , help="Force the model to be download even if already in cache-dir" ) download_parser.add_argument( "--trust-remote-code" , action="store_true" , help="Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine" , ) download_parser.add_argument("model" , type=_lowerCAmelCase , help="Name of the model to download" ) download_parser.set_defaults(func=_lowerCAmelCase ) def __init__( self: Tuple , _lowerCAmelCase: str , _lowerCAmelCase: str , _lowerCAmelCase: bool , _lowerCAmelCase: bool ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ =model UpperCAmelCase_ =cache UpperCAmelCase_ =force UpperCAmelCase_ =trust_remote_code def lowerCAmelCase__ ( self: Optional[Any] ) -> Optional[Any]: '''simple docstring''' from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
54
"""simple docstring""" def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = [0] * len(lowerCAmelCase ) UpperCAmelCase = [] UpperCAmelCase = [1] * len(lowerCAmelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowerCAmelCase ) ): if indegree[i] == 0: queue.append(lowerCAmelCase ) while queue: UpperCAmelCase = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: UpperCAmelCase = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowerCAmelCase ) print(max(lowerCAmelCase ) ) # Adjacency list of Graph lowerCAmelCase_ : str = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
673
0
def UpperCAmelCase ( a_ ) -> list: """simple docstring""" return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(a_ ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__('doctest').testmod()
55
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCamelCase_ ( a_ ): _A : Optional[int] = 'facebook/bart-large-mnli' _A : Union[str, Any] = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) _A : Dict = 'text_classifier' _A : Union[str, Any] = AutoTokenizer _A : Tuple = AutoModelForSequenceClassification _A : Optional[int] = ['text', ['text']] _A : Dict = ['text'] def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" super().setup() UpperCAmelCase = self.model.config UpperCAmelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase = int(snake_case__ ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def UpperCamelCase_ ( self , snake_case__ , snake_case__ ) -> List[Any]: """simple docstring""" UpperCAmelCase = labels return self.pre_processor( [text] * len(snake_case__ ) , [f'''This example is {label}''' for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def UpperCamelCase_ ( self , snake_case__ ) -> str: """simple docstring""" UpperCAmelCase = outputs.logits UpperCAmelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
673
0
'''simple docstring''' from __future__ import annotations from typing import Any def _a (lowercase__ : list ) -> int: """simple docstring""" if not postfix_notation: return 0 __snake_case = {'+', '-', '*', '/'} __snake_case = [] for token in postfix_notation: if token in operations: __snake_case , __snake_case = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(lowercase__ ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
56
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class UpperCamelCase_ ( a_ ): _A : Union[List[PIL.Image.Image], np.ndarray] _A : Optional[List[bool]] if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
673
0
def snake_case (UpperCAmelCase__ ) -> str: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" UpperCamelCase_: Optional[Any] = False if num < 0: UpperCamelCase_: Tuple = True UpperCamelCase_: int = -num UpperCamelCase_: list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(UpperCAmelCase__ ) for e in binary ) return "0b" + "".join(str(UpperCAmelCase__ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
57
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase_ : Any = { '''configuration_encodec''': [ '''ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EncodecConfig''', ], '''feature_extraction_encodec''': ['''EncodecFeatureExtractor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : List[str] = [ '''ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EncodecModel''', '''EncodecPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys lowerCAmelCase_ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
673
0
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __lowerCAmelCase ( __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Optional[int] , __UpperCamelCase : Optional[Any] ): '''simple docstring''' if openai_config_file == "": snake_case_ : List[Any] = OpenAIGPTConfig() else: snake_case_ : Optional[int] = OpenAIGPTConfig.from_json_file(__UpperCamelCase ) snake_case_ : List[Any] = OpenAIGPTModel(__UpperCamelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Save pytorch-model snake_case_ : List[Any] = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME snake_case_ : Tuple = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , __UpperCamelCase ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--openai_checkpoint_folder_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--openai_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained OpenAI model. \n''' '''This specifies the model architecture.''' ), ) __lowerCAmelCase : Union[str, Any] = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
58
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
673
0
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup __A = logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__(self : Optional[int] , **UpperCAmelCase_ : Optional[Any]) ->str: '''simple docstring''' requires_backends(self , ["bs4"]) super().__init__(**UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[Any] , UpperCAmelCase_ : Optional[int]) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Any =[] lowerCamelCase__: List[str] =[] lowerCamelCase__: Optional[Any] =element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag lowerCamelCase__: Any =parent.find_all(child.name , recursive=UpperCAmelCase_) xpath_tags.append(child.name) xpath_subscripts.append( 0 if 1 == len(UpperCAmelCase_) else next(i for i, s in enumerate(UpperCAmelCase_ , 1) if s is child)) lowerCamelCase__: Union[str, Any] =parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : str) ->Dict: '''simple docstring''' lowerCamelCase__: Dict =BeautifulSoup(UpperCAmelCase_ , "html.parser") lowerCamelCase__: Tuple =[] lowerCamelCase__: Tuple =[] lowerCamelCase__: Any =[] for element in html_code.descendants: if type(UpperCAmelCase_) == bsa.element.NavigableString: if type(element.parent) != bsa.element.Tag: continue lowerCamelCase__: Tuple =html.unescape(UpperCAmelCase_).strip() if not text_in_this_tag: continue all_doc_strings.append(UpperCAmelCase_) lowerCamelCase__ , lowerCamelCase__: Optional[int] =self.xpath_soup(UpperCAmelCase_) stringaxtag_seq.append(UpperCAmelCase_) stringaxsubs_seq.append(UpperCAmelCase_) if len(UpperCAmelCase_) != len(UpperCAmelCase_): raise ValueError("Number of doc strings and xtags does not correspond") if len(UpperCAmelCase_) != len(UpperCAmelCase_): raise ValueError("Number of doc strings and xsubs does not correspond") return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any]) ->Union[str, Any]: '''simple docstring''' lowerCamelCase__: List[Any] ="" for tagname, subs in zip(UpperCAmelCase_ , UpperCAmelCase_): xpath += F"""/{tagname}""" if subs != 0: xpath += F"""[{subs}]""" return xpath def __call__(self : Tuple , UpperCAmelCase_ : Any) ->BatchFeature: '''simple docstring''' lowerCamelCase__: List[str] =False # Check that strings has a valid type if isinstance(UpperCAmelCase_ , UpperCAmelCase_): lowerCamelCase__: int =True elif isinstance(UpperCAmelCase_ , (list, tuple)): if len(UpperCAmelCase_) == 0 or isinstance(html_strings[0] , UpperCAmelCase_): lowerCamelCase__: str =True if not valid_strings: raise ValueError( "HTML strings must of type `str`, `List[str]` (batch of examples), " F"""but is of type {type(UpperCAmelCase_)}.""") lowerCamelCase__: Tuple =bool(isinstance(UpperCAmelCase_ , (list, tuple)) and (isinstance(html_strings[0] , UpperCAmelCase_))) if not is_batched: lowerCamelCase__: List[Any] =[html_strings] # Get nodes + xpaths lowerCamelCase__: List[str] =[] lowerCamelCase__: List[str] =[] for html_string in html_strings: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Any =self.get_three_from_single(UpperCAmelCase_) nodes.append(UpperCAmelCase_) lowerCamelCase__: Any =[] for node, tag_list, sub_list in zip(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_): lowerCamelCase__: List[Any] =self.construct_xpath(UpperCAmelCase_ , UpperCAmelCase_) xpath_strings.append(UpperCAmelCase_) xpaths.append(UpperCAmelCase_) # return as Dict lowerCamelCase__: int ={"nodes": nodes, "xpaths": xpaths} lowerCamelCase__: int =BatchFeature(data=UpperCAmelCase_ , tensor_type=UpperCAmelCase_) return encoded_inputs
59
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ ( a_ , unittest.TestCase ): _A : str = VideoToVideoSDPipeline _A : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} _A : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} _A : int = PipelineTesterMixin.required_optional_params - {'latents'} _A : List[str] = False # No `output_type`. _A : Any = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCamelCase_ ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) UpperCAmelCase = CLIPTextModel(snake_case__ ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase_ ( self , snake_case__ , snake_case__=0 ) -> List[str]: """simple docstring""" UpperCAmelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(snake_case__ ) else: UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """video""": video, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = VideoToVideoSDPipeline(**snake_case__ ) UpperCAmelCase = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) UpperCAmelCase = """np""" UpperCAmelCase = sd_pipe(**snake_case__ ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) UpperCAmelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" pass def UpperCamelCase_ ( self ) -> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase_ ( unittest.TestCase ): def UpperCamelCase_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames UpperCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCAmelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case__ ) UpperCAmelCase = video.to("""cuda""" ) UpperCAmelCase = """Spiderman is surfing""" UpperCAmelCase = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type="""pt""" ).frames UpperCAmelCase = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
673
0