code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
def snake_case (__lowercase , __lowercase ) -> int: '''simple docstring''' while second != 0: _snake_case : List[str] = first & second first ^= second _snake_case : str = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : Any = int(input('Enter the first number: ').strip()) __SCREAMING_SNAKE_CASE : Any = int(input('Enter the second number: ').strip()) print(F'''{add(first, second) = }''')
670
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase_ : _lowerCamelCase = LEDConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=4 , ): _snake_case : Optional[int] = parent _snake_case : str = batch_size _snake_case : int = seq_length _snake_case : Dict = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = vocab_size _snake_case : str = hidden_size _snake_case : int = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : int = intermediate_size _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : int = max_position_embeddings _snake_case : Union[str, Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Any = bos_token_id _snake_case : str = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after _snake_case : List[Any] = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests _snake_case : List[str] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def UpperCamelCase ( self ): _snake_case : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Any = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) _snake_case : Optional[Any] = prepare_led_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) _snake_case : int = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]] , axis=-1 , ) _snake_case : List[Any] = global_attention_mask return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = TFLEDModel(config=lowercase_ ).get_decoder() _snake_case : Optional[Any] = inputs_dict["input_ids"] _snake_case : Optional[int] = input_ids[:1, :] _snake_case : int = inputs_dict["attention_mask"][:1, :] _snake_case : int = 1 # first forward pass _snake_case : str = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case : Tuple = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case : List[str] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case : str = model(lowercase_ , attention_mask=lowercase_ )[0] _snake_case : List[str] = model(lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_ , lowercase_ , rtol=1e-3 ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> List[Any]: '''simple docstring''' if attention_mask is None: _snake_case : int = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () _lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFLEDModelTester(self ) _snake_case : List[Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = tf.zeros_like(inputs_dict["attention_mask"] ) _snake_case : Tuple = 2 _snake_case : Dict = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) _snake_case : Tuple = True _snake_case : Union[str, Any] = self.model_tester.seq_length _snake_case : Union[str, Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ ): _snake_case : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(lowercase_ ): _snake_case : int = [t.numpy() for t in outputs.encoder_attentions] _snake_case : Optional[int] = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: _snake_case : Union[str, Any] = True _snake_case : Dict = False _snake_case : Any = False _snake_case : Any = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) _snake_case : Tuple = len(lowercase_ ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: _snake_case : int = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Any = model_class(lowercase_ ) _snake_case : Optional[Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine _snake_case : Optional[int] = True _snake_case : Optional[int] = True _snake_case : List[Any] = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # TODO: Head-masking not yet implement pass def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' return tf.constant(__lowercase , dtype=tf.intaa ) __SCREAMING_SNAKE_CASE : List[Any] = 1E-4 @slow @require_tf class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Dict = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here _snake_case : Union[str, Any] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[int] = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Union[str, Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Optional[Any] = model(**lowercase_ )[0] _snake_case : str = (1, 1_024, 768) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[Any] = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 ) def UpperCamelCase ( self ): _snake_case : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here _snake_case : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : int = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Tuple = model(**lowercase_ )[0] _snake_case : Any = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[int] = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 , rtol=1e-3 )
670
1
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py __SCREAMING_SNAKE_CASE : Union[str, Any] = 'src/transformers' __SCREAMING_SNAKE_CASE : Dict = 'docs/source/en' __SCREAMING_SNAKE_CASE : List[Any] = '.' def snake_case (__lowercase , __lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' with open(__lowercase , "r" , encoding="utf-8" , newline="\n" ) as f: _snake_case : List[Any] = f.readlines() # Find the start prompt. _snake_case : Optional[int] = 0 while not lines[start_index].startswith(__lowercase ): start_index += 1 start_index += 1 _snake_case : Optional[Any] = start_index while not lines[end_index].startswith(__lowercase ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | __SCREAMING_SNAKE_CASE : Optional[int] = 'Model|Encoder|Decoder|ForConditionalGeneration' # Regexes that match TF/Flax/PT model names. __SCREAMING_SNAKE_CASE : Tuple = re.compile(R'TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') __SCREAMING_SNAKE_CASE : Union[str, Any] = re.compile(R'Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. __SCREAMING_SNAKE_CASE : Any = re.compile(R'(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') # This is to make sure the transformers module imported is the one in the repo. __SCREAMING_SNAKE_CASE : Union[str, Any] = direct_transformers_import(TRANSFORMERS_PATH) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = re.finditer(".+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)" , __lowercase ) return [m.group(0 ) for m in matches] def snake_case (__lowercase , __lowercase ) -> Optional[Any]: '''simple docstring''' _snake_case : Any = 2 if text == "✅" or text == "❌" else len(__lowercase ) _snake_case : List[str] = (width - text_length) // 2 _snake_case : Union[str, Any] = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def snake_case () -> Dict: '''simple docstring''' _snake_case : int = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES _snake_case : Tuple = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } _snake_case : Tuple = {name: config.replace("Config" , "" ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. _snake_case : Optional[int] = collections.defaultdict(__lowercase ) _snake_case : Optional[Any] = collections.defaultdict(__lowercase ) _snake_case : Union[str, Any] = collections.defaultdict(__lowercase ) _snake_case : str = collections.defaultdict(__lowercase ) _snake_case : int = collections.defaultdict(__lowercase ) # Let's lookup through all transformers object (once). for attr_name in dir(__lowercase ): _snake_case : str = None if attr_name.endswith("Tokenizer" ): _snake_case : Union[str, Any] = slow_tokenizers _snake_case : str = attr_name[:-9] elif attr_name.endswith("TokenizerFast" ): _snake_case : Dict = fast_tokenizers _snake_case : Optional[int] = attr_name[:-13] elif _re_tf_models.match(__lowercase ) is not None: _snake_case : str = tf_models _snake_case : int = _re_tf_models.match(__lowercase ).groups()[0] elif _re_flax_models.match(__lowercase ) is not None: _snake_case : List[str] = flax_models _snake_case : Optional[int] = _re_flax_models.match(__lowercase ).groups()[0] elif _re_pt_models.match(__lowercase ) is not None: _snake_case : Dict = pt_models _snake_case : Dict = _re_pt_models.match(__lowercase ).groups()[0] if lookup_dict is not None: while len(__lowercase ) > 0: if attr_name in model_name_to_prefix.values(): _snake_case : Dict = True break # Try again after removing the last word in the name _snake_case : int = "".join(camel_case_split(__lowercase )[:-1] ) # Let's build that table! _snake_case : List[str] = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) _snake_case : List[str] = ["Model", "Tokenizer slow", "Tokenizer fast", "PyTorch support", "TensorFlow support", "Flax Support"] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). _snake_case : str = [len(__lowercase ) + 2 for c in columns] _snake_case : Optional[Any] = max([len(__lowercase ) for name in model_names] ) + 2 # Build the table per se _snake_case : str = "|" + "|".join([_center_text(__lowercase , __lowercase ) for c, w in zip(__lowercase , __lowercase )] ) + "|\n" # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([":" + "-" * (w - 2) + ":" for w in widths] ) + "|\n" _snake_case : List[Any] = {True: "✅", False: "❌"} for name in model_names: _snake_case : Optional[Any] = model_name_to_prefix[name] _snake_case : List[str] = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(__lowercase , __lowercase ) for l, w in zip(__lowercase , __lowercase )] ) + "|\n" return table def snake_case (__lowercase=False ) -> int: '''simple docstring''' _snake_case ,_snake_case ,_snake_case ,_snake_case : Any = _find_text_in_file( filename=os.path.join(__lowercase , "index.md" ) , start_prompt="<!--This table is updated automatically from the auto modules" , end_prompt="<!-- End table-->" , ) _snake_case : List[str] = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(__lowercase , "index.md" ) , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( "The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this." ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') __SCREAMING_SNAKE_CASE : Any = parser.parse_args() check_model_table(args.fix_and_overwrite)
670
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
1
import json import os import torch from diffusers import UNetaDModel os.makedirs('hub/hopper-medium-v2/unet/hor32', exist_ok=True) os.makedirs('hub/hopper-medium-v2/unet/hor128', exist_ok=True) os.makedirs('hub/hopper-medium-v2/value_function', exist_ok=True) def snake_case (__lowercase ) -> Any: '''simple docstring''' if hor == 128: _snake_case : List[str] = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") _snake_case : Optional[Any] = (32, 128, 256) _snake_case : Optional[Any] = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: _snake_case : int = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") _snake_case : Optional[Any] = (32, 64, 128, 256) _snake_case : str = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") _snake_case : Tuple = torch.load(F"""/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch""" ) _snake_case : Dict = model.state_dict() _snake_case : str = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 65_536, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } _snake_case : str = UNetaDModel(**__lowercase ) print(F"""length of state dict: {len(state_dict.keys() )}""" ) print(F"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) _snake_case : Optional[int] = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _snake_case : Optional[Any] = state_dict.pop(__lowercase ) hf_value_function.load_state_dict(__lowercase ) torch.save(hf_value_function.state_dict() , F"""hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin""" ) with open(F"""hub/hopper-medium-v2/unet/hor{hor}/config.json""" , "w" ) as f: json.dump(__lowercase , __lowercase ) def snake_case () -> str: '''simple docstring''' _snake_case : List[Any] = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 128, 256), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 65_536, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } _snake_case : Optional[int] = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) _snake_case : Dict = model _snake_case : Any = UNetaDModel(**__lowercase ) print(F"""length of state dict: {len(state_dict.keys() )}""" ) print(F"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) _snake_case : Optional[Any] = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _snake_case : Tuple = state_dict.pop(__lowercase ) hf_value_function.load_state_dict(__lowercase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(__lowercase , __lowercase ) if __name__ == "__main__": unet(3_2) # unet(128) value_function()
670
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Any = tempfile.mkdtemp() # fmt: off _snake_case : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _snake_case : Dict = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] _snake_case : Optional[int] = {"unk_token": "<unk>"} _snake_case : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) _snake_case : Any = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } _snake_case : Optional[Any] = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Union[str, Any] = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Tuple = self.get_tokenizer() _snake_case : Any = self.get_rust_tokenizer() _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) _snake_case : Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) _snake_case : List[Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : List[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Tuple = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : int = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Optional[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : str = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : Dict = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : int = processor(text=lowercase_ ) _snake_case : str = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[Any] = "lower newer" _snake_case : int = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[str] = self.get_tokenizer() _snake_case : Union[str, Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Dict = self.prepare_image_inputs() _snake_case : List[Any] = processor(images=lowercase_ , visual_prompt=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[Any] = self.get_tokenizer() _snake_case : str = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Any = processor.batch_decode(lowercase_ ) _snake_case : Any = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ )
670
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class lowercase_ : def __init__( self , lowercase_ , lowercase_=2 , lowercase_=True , lowercase_=False , lowercase_=10 , lowercase_=3 , lowercase_=32 * 4 , lowercase_=32 * 6 , lowercase_=4 , lowercase_=32 , ): _snake_case : int = parent _snake_case : Dict = batch_size _snake_case : str = is_training _snake_case : Optional[int] = use_auxiliary_loss _snake_case : Optional[int] = num_queries _snake_case : Optional[Any] = num_channels _snake_case : List[Any] = min_size _snake_case : Tuple = max_size _snake_case : Tuple = num_labels _snake_case : Dict = mask_feature_size def UpperCamelCase ( self ): _snake_case : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowercase_ ) _snake_case : Any = torch.ones([self.batch_size, self.min_size, self.max_size] , device=lowercase_ ) _snake_case : Any = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=lowercase_ ) > 0.5 ).float() _snake_case : Dict = (torch.rand((self.batch_size, self.num_labels) , device=lowercase_ ) > 0.5).long() _snake_case : str = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def UpperCamelCase ( self ): return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def UpperCamelCase ( self ): _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : Optional[Any] = self.prepare_config_and_inputs() _snake_case : Dict = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = output.encoder_hidden_states _snake_case : Optional[int] = output.pixel_decoder_hidden_states _snake_case : int = output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowercase_ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowercase_ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowercase_ ) , config.decoder_config.decoder_layers ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=False ): with torch.no_grad(): _snake_case : Optional[int] = MaskFormerModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() _snake_case : Optional[Any] = model(pixel_values=lowercase_ , pixel_mask=lowercase_ ) _snake_case : List[str] = model(lowercase_ , output_hidden_states=lowercase_ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): _snake_case : str = MaskFormerForInstanceSegmentation(config=lowercase_ ) model.to(lowercase_ ) model.eval() def comm_check_on_output(lowercase_ ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _snake_case : Any = model(pixel_values=lowercase_ , pixel_mask=lowercase_ ) _snake_case : Optional[int] = model(lowercase_ ) comm_check_on_output(lowercase_ ) _snake_case : Dict = model( pixel_values=lowercase_ , pixel_mask=lowercase_ , mask_labels=lowercase_ , class_labels=lowercase_ ) comm_check_on_output(lowercase_ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () _lowerCamelCase = ( {'feature-extraction': MaskFormerModel, 'image-segmentation': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[int] = MaskFormerModelTester(self ) _snake_case : Tuple = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowercase_ , **lowercase_ , output_hidden_states=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*lowercase_ ) @unittest.skip(reason="MaskFormer does not use inputs_embeds" ) def UpperCamelCase ( self ): pass @unittest.skip(reason="MaskFormer does not have a get_input_embeddings method" ) def UpperCamelCase ( self ): pass @unittest.skip(reason="MaskFormer is not a generative model" ) def UpperCamelCase ( self ): pass @unittest.skip(reason="MaskFormer does not use token embeddings" ) def UpperCamelCase ( self ): pass @require_torch_multi_gpu @unittest.skip( reason="MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def UpperCamelCase ( self ): pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Tuple = model_class(lowercase_ ) _snake_case : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : Tuple = [*signature.parameters.keys()] _snake_case : Union[str, Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_ ) @slow def UpperCamelCase ( self ): for model_name in ["facebook/maskformer-swin-small-coco"]: _snake_case : Optional[Any] = MaskFormerModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = (self.model_tester.min_size,) * 2 _snake_case : List[str] = { "pixel_values": torch.randn((2, 3, *size) , device=lowercase_ ), "mask_labels": torch.randn((2, 10, *size) , device=lowercase_ ), "class_labels": torch.zeros(2 , 10 , device=lowercase_ ).long(), } _snake_case : Any = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(lowercase_ ) _snake_case : Optional[int] = model(**lowercase_ ) self.assertTrue(outputs.loss is not None ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowercase_ , **lowercase_ , output_hidden_states=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Union[str, Any] = model_class(lowercase_ ).to(lowercase_ ) _snake_case : Any = model(**lowercase_ , output_attentions=lowercase_ ) self.assertTrue(outputs.attentions is not None ) def UpperCamelCase ( self ): if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _snake_case : Union[str, Any] = self.all_model_classes[1] _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() _snake_case : Tuple = model_class(lowercase_ ) model.to(lowercase_ ) model.train() _snake_case : Optional[Any] = model(lowercase_ , mask_labels=lowercase_ , class_labels=lowercase_ ).loss loss.backward() def UpperCamelCase ( self ): # only MaskFormerForInstanceSegmentation has the loss _snake_case : Union[str, Any] = self.all_model_classes[1] _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : Tuple = self.model_tester.prepare_config_and_inputs() _snake_case : List[Any] = True _snake_case : str = True _snake_case : Dict = model_class(lowercase_ ) model.to(lowercase_ ) model.train() _snake_case : Optional[int] = model(lowercase_ , mask_labels=lowercase_ , class_labels=lowercase_ ) _snake_case : List[str] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _snake_case : Optional[Any] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _snake_case : Any = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _snake_case : Optional[Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowercase_ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __SCREAMING_SNAKE_CASE : Any = 1E-4 def snake_case () -> List[Any]: '''simple docstring''' _snake_case : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class lowercase_ ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ): return ( MaskFormerImageProcessor.from_pretrained("facebook/maskformer-swin-small-coco" ) if is_vision_available() else None ) def UpperCamelCase ( self ): _snake_case : Dict = MaskFormerModel.from_pretrained("facebook/maskformer-swin-small-coco" ).to(lowercase_ ) _snake_case : Dict = self.default_image_processor _snake_case : List[Any] = prepare_img() _snake_case : Tuple = image_processor(lowercase_ , return_tensors="pt" ).to(lowercase_ ) _snake_case : Union[str, Any] = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowercase_ , (1, 3, 800, 1_088) ) with torch.no_grad(): _snake_case : int = model(**lowercase_ ) _snake_case : Optional[int] = torch.tensor( [[-0.0_482, 0.9_228, 0.4_951], [-0.2_547, 0.8_017, 0.8_527], [-0.0_069, 0.3_385, -0.0_089]] ).to(lowercase_ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , lowercase_ , atol=lowercase_ ) ) _snake_case : List[Any] = torch.tensor( [[-0.8_422, -0.8_434, -0.9_718], [-1.0_144, -0.5_565, -0.4_195], [-1.0_038, -0.4_484, -0.1_961]] ).to(lowercase_ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , lowercase_ , atol=lowercase_ ) ) _snake_case : List[Any] = torch.tensor( [[0.2_852, -0.0_159, 0.9_735], [0.6_254, 0.1_858, 0.8_529], [-0.0_680, -0.4_116, 1.8_413]] ).to(lowercase_ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , lowercase_ , atol=lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : List[Any] = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(lowercase_ ) .eval() ) _snake_case : List[str] = self.default_image_processor _snake_case : Any = prepare_img() _snake_case : str = image_processor(lowercase_ , return_tensors="pt" ).to(lowercase_ ) _snake_case : List[Any] = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowercase_ , (1, 3, 800, 1_088) ) with torch.no_grad(): _snake_case : Optional[Any] = model(**lowercase_ ) # masks_queries_logits _snake_case : Optional[int] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case : List[Any] = [ [-1.3_737_124, -1.7_724_937, -1.9_364_233], [-1.5_977_281, -1.9_867_939, -2.1_523_695], [-1.5_795_398, -1.9_269_832, -2.093_942], ] _snake_case : Union[str, Any] = torch.tensor(lowercase_ ).to(lowercase_ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowercase_ , atol=lowercase_ ) ) # class_queries_logits _snake_case : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case : Dict = torch.tensor( [ [1.65_12e00, -5.25_72e00, -3.35_19e00], [3.61_69e-02, -5.90_25e00, -2.93_13e00], [1.07_66e-04, -7.76_30e00, -5.12_63e00], ] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowercase_ , atol=lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : Dict = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-resnet101-coco-stuff" ) .to(lowercase_ ) .eval() ) _snake_case : Tuple = self.default_image_processor _snake_case : Tuple = prepare_img() _snake_case : Any = image_processor(lowercase_ , return_tensors="pt" ).to(lowercase_ ) _snake_case : Dict = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowercase_ , (1, 3, 800, 1_088) ) with torch.no_grad(): _snake_case : int = model(**lowercase_ ) # masks_queries_logits _snake_case : Any = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case : Optional[int] = [[-0.9_046, -2.6_366, -4.6_062], [-3.4_179, -5.7_890, -8.8_057], [-4.9_179, -7.6_560, -10.7_711]] _snake_case : Optional[int] = torch.tensor(lowercase_ ).to(lowercase_ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowercase_ , atol=lowercase_ ) ) # class_queries_logits _snake_case : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case : List[str] = torch.tensor( [[4.7_188, -3.2_585, -2.8_857], [6.6_871, -2.9_181, -1.2_487], [7.2_449, -2.2_764, -2.1_874]] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowercase_ , atol=lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : Any = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(lowercase_ ) .eval() ) _snake_case : int = self.default_image_processor _snake_case : List[Any] = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) _snake_case : Dict = inputs["pixel_values"].to(lowercase_ ) _snake_case : Optional[Any] = [el.to(lowercase_ ) for el in inputs["mask_labels"]] _snake_case : Any = [el.to(lowercase_ ) for el in inputs["class_labels"]] with torch.no_grad(): _snake_case : int = model(**lowercase_ ) self.assertTrue(outputs.loss is not None )
670
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
1
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: _snake_case : List[Any] = ( f"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" f""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowercase_ , standard_warn=lowercase_ ) _snake_case : Any = dict(scheduler.config ) _snake_case : int = 1 _snake_case : List[str] = FrozenDict(lowercase_ ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: _snake_case : Optional[int] = ( f"""The configuration file of this scheduler: {scheduler} has not set the configuration""" " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowercase_ , standard_warn=lowercase_ ) _snake_case : Dict = dict(scheduler.config ) _snake_case : Tuple = True _snake_case : str = FrozenDict(lowercase_ ) if safety_checker is None: logger.warning( f"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowercase_ , segmentation_processor=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , unet=lowercase_ , scheduler=lowercase_ , safety_checker=lowercase_ , feature_extractor=lowercase_ , ) def UpperCamelCase ( self , lowercase_ = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory _snake_case : Tuple = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase_ ) def UpperCamelCase ( self ): self.enable_attention_slicing(lowercase_ ) def UpperCamelCase ( self ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) _snake_case : int = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase ( self ): if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 50 , lowercase_ = 7.5 , lowercase_ = None , lowercase_ = 1 , lowercase_ = 0.0 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , lowercase_ = None , lowercase_ = 1 , **lowercase_ , ): _snake_case : int = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) _snake_case : Optional[int] = self.segmentation_model(**lowercase_ ) _snake_case : List[Any] = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() _snake_case : Optional[int] = self.numpy_to_pil(lowercase_ )[0].resize(image.size ) # Run inpainting pipeline with the generated mask _snake_case : List[Any] = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowercase_ , image=lowercase_ , mask_image=lowercase_ , height=lowercase_ , width=lowercase_ , num_inference_steps=lowercase_ , guidance_scale=lowercase_ , negative_prompt=lowercase_ , num_images_per_prompt=lowercase_ , eta=lowercase_ , generator=lowercase_ , latents=lowercase_ , output_type=lowercase_ , return_dict=lowercase_ , callback=lowercase_ , callback_steps=lowercase_ , )
670
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ ): super().__init__() _snake_case : List[str] = nn.ModuleList(lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = True , ): for i, (image, scale, controlnet) in enumerate(zip(lowercase_ , lowercase_ , self.nets ) ): _snake_case ,_snake_case : Optional[int] = controlnet( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) # merge samples if i == 0: _snake_case ,_snake_case : Tuple = down_samples, mid_sample else: _snake_case : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase_ , lowercase_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , ): _snake_case : Tuple = 0 _snake_case : Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase_ , is_main_process=lowercase_ , save_function=lowercase_ , safe_serialization=lowercase_ , variant=lowercase_ , ) idx += 1 _snake_case : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : List[str] = 0 _snake_case : Optional[Any] = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case : Optional[Any] = pretrained_model_path while os.path.isdir(lowercase_ ): _snake_case : int = ControlNetModel.from_pretrained(lowercase_ , **lowercase_ ) controlnets.append(lowercase_ ) idx += 1 _snake_case : str = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(lowercase_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(lowercase_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(lowercase_ )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(lowercase_ )
670
1
def snake_case (__lowercase ) -> int: '''simple docstring''' _snake_case : List[str] = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case (__lowercase ) -> int: '''simple docstring''' _snake_case : Union[str, Any] = 0 while number > 0: _snake_case : Optional[Any] = number % 10 sum_of_digits += last_digit _snake_case : Dict = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case (__lowercase = 100 ) -> int: '''simple docstring''' _snake_case : str = factorial(__lowercase ) _snake_case : Dict = split_and_add(__lowercase ) return result if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
670
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'CLIPImageProcessor' _lowerCamelCase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Dict = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _snake_case : str = self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if images is not None: _snake_case : List[str] = self.image_processor(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if text is not None and images is not None: _snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase_ ) , tensor_type=lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
1
import itertools import string from collections.abc import Generator, Iterable def snake_case (__lowercase , __lowercase ) -> Generator[tuple[str, ...], None, None]: '''simple docstring''' _snake_case : int = iter(__lowercase ) while True: _snake_case : Tuple = tuple(itertools.islice(__lowercase , __lowercase ) ) if not chunk: return yield chunk def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "".join([c.upper() for c in dirty if c in string.ascii_letters] ) _snake_case : Union[str, Any] = "" if len(__lowercase ) < 2: return dirty for i in range(len(__lowercase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(__lowercase ) & 1: clean += "X" return clean def snake_case (__lowercase ) -> list[str]: '''simple docstring''' _snake_case : Any = "ABCDEFGHIKLMNOPQRSTUVWXYZ" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _snake_case : int = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(__lowercase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(__lowercase ) return table def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = generate_table(__lowercase ) _snake_case : List[str] = prepare_input(__lowercase ) _snake_case : Tuple = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowercase , 2 ): _snake_case ,_snake_case : Tuple = divmod(table.index(__lowercase ) , 5 ) _snake_case ,_snake_case : Optional[Any] = divmod(table.index(__lowercase ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = generate_table(__lowercase ) _snake_case : List[str] = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowercase , 2 ): _snake_case ,_snake_case : Tuple = divmod(table.index(__lowercase ) , 5 ) _snake_case ,_snake_case : Dict = divmod(table.index(__lowercase ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
670
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
670
1
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'Pix2StructImageProcessor' _lowerCamelCase = ('T5Tokenizer', 'T5TokenizerFast') def __init__( self , lowercase_ , lowercase_ ): _snake_case : Tuple = False super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_ = None , lowercase_ = True , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = 2_048 , lowercase_ = 0 , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = True , lowercase_ = None , **lowercase_ , ): if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None and not self.image_processor.is_vqa: _snake_case : List[Any] = self.tokenizer _snake_case : int = self.tokenizer( text=lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , stride=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , return_overflowing_tokens=lowercase_ , return_special_tokens_mask=lowercase_ , return_offsets_mapping=lowercase_ , return_token_type_ids=lowercase_ , return_length=lowercase_ , verbose=lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values _snake_case : Dict = self.image_processor( lowercase_ , return_tensors=lowercase_ , max_patches=lowercase_ , **lowercase_ ) else: # add pixel_values and bbox _snake_case : Optional[Any] = self.image_processor( lowercase_ , return_tensors=lowercase_ , max_patches=lowercase_ , header_text=lowercase_ , **lowercase_ ) if text is not None and not self.image_processor.is_vqa: _snake_case : Union[str, Any] = self.tokenizer( text=lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , stride=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , return_overflowing_tokens=lowercase_ , return_special_tokens_mask=lowercase_ , return_offsets_mapping=lowercase_ , return_token_type_ids=lowercase_ , return_length=lowercase_ , verbose=lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) if "attention_mask" in text_encoding: _snake_case : str = text_encoding.pop("attention_mask" ) if "input_ids" in text_encoding: _snake_case : List[str] = text_encoding.pop("input_ids" ) else: _snake_case : Dict = None if text_encoding is not None: encoding_image_processor.update(lowercase_ ) return encoding_image_processor def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : Optional[int] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def snake_case (*__lowercase ) -> Dict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): _snake_case : Dict = list(__lowercase ) for i in range(len(__lowercase ) ): _snake_case : List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : str = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowercase , __lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def snake_case (__lowercase = None , __lowercase = 128 ) -> Any: '''simple docstring''' if function is None: return functools.partial(__lowercase , starting_batch_size=__lowercase ) _snake_case : List[str] = starting_batch_size def decorator(*__lowercase , **__lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _snake_case : Optional[Any] = list(inspect.signature(__lowercase ).parameters.keys() ) # Guard against user error if len(__lowercase ) < (len(__lowercase ) + 1): _snake_case : str = ", ".join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowercase , *__lowercase , **__lowercase ) except Exception as e: if should_reduce_batch_size(__lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
670
1
__SCREAMING_SNAKE_CASE : Any = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' __SCREAMING_SNAKE_CASE : int = [{'type': 'code', 'content': INSTALL_CONTENT}] __SCREAMING_SNAKE_CASE : List[str] = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
670
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
1
import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_=None , **lowercase_ ): warnings.warn( "`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` " "instead." , lowercase_ , ) super().__init__(args=lowercase_ , **lowercase_ )
670
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
1
import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = DownBlockaD # noqa F405 _lowerCamelCase = 'down' def UpperCamelCase ( self ): _snake_case : Optional[Any] = [-0.0_232, -0.9_869, 0.8_054, -0.0_637, -0.1_688, -1.4_264, 0.4_470, -1.3_394, 0.0_904] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ResnetDownsampleBlockaD # noqa F405 _lowerCamelCase = 'down' def UpperCamelCase ( self ): _snake_case : int = [0.0_710, 0.2_410, -0.7_320, -1.0_757, -1.1_343, 0.3_540, -0.0_133, -0.2_576, 0.0_948] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnDownBlockaD # noqa F405 _lowerCamelCase = 'down' def UpperCamelCase ( self ): _snake_case : List[Any] = [0.0_636, 0.8_964, -0.6_234, -1.0_131, 0.0_844, 0.4_935, 0.3_437, 0.0_911, -0.2_957] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = CrossAttnDownBlockaD # noqa F405 _lowerCamelCase = 'down' def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = super().prepare_init_args_and_inputs_for_common() _snake_case : Optional[Any] = 32 return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Dict = [0.2_238, -0.7_396, -0.2_255, -0.3_829, 0.1_925, 1.1_665, 0.0_603, -0.7_295, 0.1_983] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = SimpleCrossAttnDownBlockaD # noqa F405 _lowerCamelCase = 'down' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_encoder_hidden_states=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[str] = super().prepare_init_args_and_inputs_for_common() _snake_case : Optional[int] = 32 return init_dict, inputs_dict @unittest.skipIf(torch_device == "mps" , "MPS result is not consistent" ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = [0.7_921, -0.0_992, -0.1_962, -0.7_695, -0.4_242, 0.7_804, 0.4_737, 0.2_765, 0.3_338] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = SkipDownBlockaD # noqa F405 _lowerCamelCase = 'down' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_skip_sample=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Optional[int] = [-0.0_845, -0.2_087, -0.2_465, 0.0_971, 0.1_900, -0.0_484, 0.2_664, 0.4_179, 0.5_069] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnSkipDownBlockaD # noqa F405 _lowerCamelCase = 'down' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_skip_sample=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Dict = [0.5_539, 0.1_609, 0.4_924, 0.0_537, -0.1_995, 0.4_050, 0.0_979, -0.2_721, -0.0_642] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = DownEncoderBlockaD # noqa F405 _lowerCamelCase = 'down' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_temb=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Optional[int] = { "in_channels": 32, "out_channels": 32, } _snake_case : str = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Optional[int] = [1.1_102, 0.5_302, 0.4_872, -0.0_023, -0.8_042, 0.0_483, -0.3_489, -0.5_632, 0.7_626] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnDownEncoderBlockaD # noqa F405 _lowerCamelCase = 'down' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_temb=lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = { "in_channels": 32, "out_channels": 32, } _snake_case : int = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Any = [0.8_966, -0.1_486, 0.8_568, 0.8_141, -0.9_046, -0.1_342, -0.0_972, -0.7_417, 0.1_538] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = UNetMidBlockaD # noqa F405 _lowerCamelCase = 'mid' def UpperCamelCase ( self ): _snake_case : int = { "in_channels": 32, "temb_channels": 128, } _snake_case : List[Any] = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Any = [-0.1_062, 1.7_248, 0.3_494, 1.4_569, -0.0_910, -1.2_421, -0.9_984, 0.6_736, 1.0_028] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = UNetMidBlockaDCrossAttn # noqa F405 _lowerCamelCase = 'mid' def UpperCamelCase ( self ): _snake_case ,_snake_case : str = super().prepare_init_args_and_inputs_for_common() _snake_case : Tuple = 32 return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Dict = [0.0_187, 2.4_220, 0.4_484, 1.1_203, -0.6_121, -1.5_122, -0.8_270, 0.7_851, 1.8_335] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = UNetMidBlockaDSimpleCrossAttn # noqa F405 _lowerCamelCase = 'mid' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_encoder_hidden_states=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Any = super().prepare_init_args_and_inputs_for_common() _snake_case : int = 32 return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : str = [0.7_143, 1.9_974, 0.5_448, 1.3_977, 0.1_282, -1.1_237, -1.4_238, 0.5_530, 0.8_880] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = UpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Dict = [-0.2_041, -0.4_165, -0.3_022, 0.0_041, -0.6_628, -0.7_053, 0.1_928, -0.0_325, 0.0_523] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ResnetUpsampleBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[str] = [0.2_287, 0.3_549, -0.1_346, 0.4_797, -0.1_715, -0.9_649, 0.7_305, -0.5_864, -0.6_244] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = CrossAttnUpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : int = super().prepare_init_args_and_inputs_for_common() _snake_case : Optional[int] = 32 return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : str = [-0.1_403, -0.3_515, -0.0_420, -0.1_425, 0.3_167, 0.5_094, -0.2_181, 0.5_931, 0.5_582] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = SimpleCrossAttnUpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ , include_encoder_hidden_states=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : str = super().prepare_init_args_and_inputs_for_common() _snake_case : Optional[Any] = 32 return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Optional[int] = [0.2_645, 0.1_480, 0.0_909, 0.8_044, -0.9_758, -0.9_083, 0.0_994, -1.1_453, -0.7_402] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnUpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) @unittest.skipIf(torch_device == "mps" , "MPS result is not consistent" ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = [0.0_979, 0.1_326, 0.0_021, 0.0_659, 0.2_249, 0.0_059, 0.1_132, 0.5_952, 0.1_033] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = SkipUpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = [-0.0_893, -0.1_234, -0.1_506, -0.0_332, 0.0_123, -0.0_211, 0.0_566, 0.0_143, 0.0_362] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnSkipUpBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Dict = [0.0_361, 0.0_617, 0.2_787, -0.0_350, 0.0_342, 0.3_421, -0.0_843, 0.0_913, 0.3_015] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = UpDecoderBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_temb=lowercase_ ) def UpperCamelCase ( self ): _snake_case : int = {"in_channels": 32, "out_channels": 32} _snake_case : Dict = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : str = [0.4_404, 0.1_998, -0.9_886, -0.3_320, -0.3_128, -0.7_034, -0.6_955, -0.2_338, -0.3_137] super().test_output(lowercase_ ) class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = AttnUpDecoderBlockaD # noqa F405 _lowerCamelCase = 'up' @property def UpperCamelCase ( self ): return super().get_dummy_input(include_temb=lowercase_ ) def UpperCamelCase ( self ): _snake_case : Any = {"in_channels": 32, "out_channels": 32} _snake_case : Optional[int] = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ): _snake_case : Dict = [0.6_738, 0.4_491, 0.1_055, 1.0_710, 0.7_316, 0.3_339, 0.3_352, 0.1_023, 0.3_568] super().test_output(lowercase_ )
670
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : Tuple = Rectangle(height=0.5 , width=0.5 ) _snake_case : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : List[str] = [mem.copy() for i in range(6 )] _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("CPU" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : int = [mem.copy() for i in range(4 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = Text("GPU" , font_size=24 ) _snake_case : Optional[int] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = Text("Model" , font_size=24 ) _snake_case : Dict = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case : Union[str, Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : Union[str, Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Union[str, Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Optional[Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _snake_case : List[Any] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : int = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Dict = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) _snake_case : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
670
1
from collections import defaultdict def snake_case (__lowercase ) -> int: '''simple docstring''' _snake_case : int = 1 _snake_case : Union[str, Any] = True for v in tree[start]: if v not in visited: ret += dfs(__lowercase ) if ret % 2 == 0: cuts.append(__lowercase ) return ret def snake_case () -> Union[str, Any]: '''simple docstring''' dfs(1 ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE : Optional[Any] = 1_0, 9 __SCREAMING_SNAKE_CASE : List[Any] = defaultdict(list) __SCREAMING_SNAKE_CASE : dict[int, bool] = {} __SCREAMING_SNAKE_CASE : list[int] = [] __SCREAMING_SNAKE_CASE : List[Any] = 0 __SCREAMING_SNAKE_CASE : int = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (1_0, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
670
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
1
import sys import turtle def snake_case (__lowercase , __lowercase ) -> tuple[float, float]: '''simple docstring''' return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def snake_case (__lowercase , __lowercase , __lowercase , __lowercase , ) -> None: '''simple docstring''' my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(__lowercase , get_mid(__lowercase , __lowercase ) , get_mid(__lowercase , __lowercase ) , depth - 1 ) triangle(__lowercase , get_mid(__lowercase , __lowercase ) , get_mid(__lowercase , __lowercase ) , depth - 1 ) triangle(__lowercase , get_mid(__lowercase , __lowercase ) , get_mid(__lowercase , __lowercase ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( 'Correct format for using this script: ' 'python fractals.py <int:depth_for_fractal>' ) __SCREAMING_SNAKE_CASE : Union[str, Any] = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor('red') __SCREAMING_SNAKE_CASE : Optional[int] = [(-1_7_5, -1_2_5), (0, 1_7_5), (1_7_5, -1_2_5)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
670
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowercase_ ( __snake_case ): _lowerCamelCase = 'roc_bert' def __init__( self , lowercase_=30_522 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3_072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1e-12 , lowercase_=True , lowercase_=0 , lowercase_="absolute" , lowercase_=None , lowercase_=True , lowercase_=True , lowercase_=768 , lowercase_=910 , lowercase_=512 , lowercase_=24_858 , lowercase_=True , **lowercase_ , ): _snake_case : int = vocab_size _snake_case : Union[str, Any] = max_position_embeddings _snake_case : Union[str, Any] = hidden_size _snake_case : Dict = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : List[Any] = hidden_act _snake_case : Optional[int] = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : List[Any] = type_vocab_size _snake_case : int = layer_norm_eps _snake_case : Optional[Any] = use_cache _snake_case : List[Any] = enable_pronunciation _snake_case : Dict = enable_shape _snake_case : Dict = pronunciation_embed_dim _snake_case : Tuple = pronunciation_vocab_size _snake_case : Tuple = shape_embed_dim _snake_case : List[str] = shape_vocab_size _snake_case : Dict = concat_input _snake_case : int = position_embedding_type _snake_case : int = classifier_dropout super().__init__(pad_token_id=lowercase_ , **lowercase_ )
670
1
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
from cva import destroyAllWindows, imread, imshow, waitKey def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case ,_snake_case : int = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(__lowercase ): for j in range(__lowercase ): _snake_case : Optional[Any] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image __SCREAMING_SNAKE_CASE : Optional[Any] = imread('image_data/lena.jpg', 1) # convert to its negative __SCREAMING_SNAKE_CASE : Tuple = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
670
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Optional[Any] = tempfile.mkdtemp() _snake_case : Any = BlipImageProcessor() _snake_case : Optional[int] = GPTaTokenizer.from_pretrained("hf-internal-testing/tiny-random-GPT2Model" ) _snake_case : int = BlipaProcessor(lowercase_ , lowercase_ ) processor.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self , **lowercase_ ): return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase_ ).tokenizer def UpperCamelCase ( self , **lowercase_ ): return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase_ ).image_processor def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : Dict = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Dict = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Optional[int] = BlipaProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Union[str, Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Any = BlipaProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Union[str, Any] = self.get_tokenizer() _snake_case : int = BlipaProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : str = self.prepare_image_inputs() _snake_case : List[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : List[str] = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Tuple = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = BlipaProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Any = "lower newer" _snake_case : Optional[int] = processor(text=lowercase_ ) _snake_case : Optional[Any] = tokenizer(lowercase_ , return_token_type_ids=lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Tuple = self.get_tokenizer() _snake_case : int = BlipaProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Tuple = "lower newer" _snake_case : List[Any] = self.prepare_image_inputs() _snake_case : Optional[Any] = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : str = self.get_image_processor() _snake_case : Optional[int] = self.get_tokenizer() _snake_case : List[str] = BlipaProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Union[str, Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Union[str, Any] = processor.batch_decode(lowercase_ ) _snake_case : Optional[Any] = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Dict = self.get_tokenizer() _snake_case : List[Any] = BlipaProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : str = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] )
670
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __SCREAMING_SNAKE_CASE : List[str] = Mapping[str, np.ndarray] __SCREAMING_SNAKE_CASE : List[Any] = Mapping[str, Any] # Is a nested dict. __SCREAMING_SNAKE_CASE : List[Any] = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class lowercase_ : _lowerCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _lowerCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _lowerCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _lowerCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _lowerCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _lowerCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _lowerCamelCase = None # Templates used to generate this protein (prediction-only) _lowerCamelCase = None # Chain corresponding to each parent _lowerCamelCase = None def snake_case (__lowercase ) -> Protein: '''simple docstring''' _snake_case : str = r"(\[[A-Z]+\]\n)" _snake_case : List[str] = [tag.strip() for tag in re.split(__lowercase , __lowercase ) if len(__lowercase ) > 0] _snake_case : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("\n" ) for l in tags[1::2]] ) _snake_case : List[str] = ["N", "CA", "C"] _snake_case : Any = None _snake_case : Union[str, Any] = None _snake_case : Optional[int] = None for g in groups: if "[PRIMARY]" == g[0]: _snake_case : Tuple = g[1][0].strip() for i in range(len(__lowercase ) ): if seq[i] not in residue_constants.restypes: _snake_case : Tuple = "X" # FIXME: strings are immutable _snake_case : int = np.array( [residue_constants.restype_order.get(__lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: _snake_case : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(__lowercase , g[1][axis].split() ) ) ) _snake_case : Dict = np.array(__lowercase ) _snake_case : Dict = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: _snake_case : int = np.array(list(map({"-": 0, "+": 1}.get , g[1][0].strip() ) ) ) _snake_case : Any = np.zeros( ( len(__lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : Dict = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__lowercase , atom_mask=__lowercase , aatype=__lowercase , residue_index=np.arange(len(__lowercase ) ) , b_factors=__lowercase , ) def snake_case (__lowercase , __lowercase = 0 ) -> List[str]: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[Any] = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) _snake_case : str = prot.parents _snake_case : str = prot.parents_chain_index if parents is not None and parents_chain_index is not None: _snake_case : int = [p for i, p in zip(__lowercase , __lowercase ) if i == chain_id] if parents is None or len(__lowercase ) == 0: _snake_case : Optional[int] = ["N/A"] pdb_headers.append(F"""PARENT {' '.join(__lowercase )}""" ) return pdb_headers def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[int] = pdb_str.split("\n" ) _snake_case : List[str] = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) _snake_case : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: _snake_case : str = [] if prot.parents_chain_index is not None: _snake_case : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(__lowercase ) , [] ) parent_dict[str(__lowercase )].append(__lowercase ) _snake_case : Any = max([int(__lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): _snake_case : Tuple = parent_dict.get(str(__lowercase ) , ["N/A"] ) parents_per_chain.append(__lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: _snake_case : List[str] = [["N/A"]] def make_parent_line(__lowercase ) -> str: return F"""PARENT {' '.join(__lowercase )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) _snake_case : int = 0 for i, l in enumerate(__lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__lowercase ): _snake_case : Tuple = parents_per_chain[chain_counter] else: _snake_case : str = ["N/A"] out_pdb_lines.append(make_parent_line(__lowercase ) ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = residue_constants.restypes + ["X"] def res_atoa(__lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , "UNK" ) _snake_case : Optional[int] = residue_constants.atom_types _snake_case : List[str] = [] _snake_case : Tuple = prot.atom_mask _snake_case : List[str] = prot.aatype _snake_case : int = prot.atom_positions _snake_case : int = prot.residue_index.astype(np.intaa ) _snake_case : List[Any] = prot.b_factors _snake_case : str = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("Invalid aatypes." ) _snake_case : Union[str, Any] = get_pdb_headers(__lowercase ) if len(__lowercase ) > 0: pdb_lines.extend(__lowercase ) _snake_case : Optional[Any] = aatype.shape[0] _snake_case : str = 1 _snake_case : Tuple = 0 _snake_case : int = string.ascii_uppercase _snake_case : Optional[Any] = None # Add all atom sites. for i in range(__lowercase ): _snake_case : Dict = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue _snake_case : List[Any] = "ATOM" _snake_case : Union[str, Any] = atom_name if len(__lowercase ) == 4 else F""" {atom_name}""" _snake_case : str = "" _snake_case : str = "" _snake_case : Any = 1.00 _snake_case : str = atom_name[0] # Protein supports only C, N, O, S, this works. _snake_case : Dict = "" _snake_case : Any = "A" if chain_index is not None: _snake_case : List[Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! _snake_case : Optional[int] = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(__lowercase ) atom_index += 1 _snake_case : Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: _snake_case : Optional[int] = True _snake_case : Union[str, Any] = chain_index[i + 1] if should_terminate: # Close the chain. _snake_case : List[str] = "TER" _snake_case : str = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(__lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__lowercase , __lowercase ) ) pdb_lines.append("END" ) pdb_lines.append("" ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Protein: '''simple docstring''' return Protein( aatype=features["aatype"] , atom_positions=result["final_atom_positions"] , atom_mask=result["final_atom_mask"] , residue_index=features["residue_index"] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["final_atom_mask"] ) , chain_index=__lowercase , remark=__lowercase , parents=__lowercase , parents_chain_index=__lowercase , )
670
1
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor'] _lowerCamelCase = 'SamImageProcessor' def __init__( self , lowercase_ ): super().__init__(lowercase_ ) _snake_case : Optional[Any] = self.image_processor _snake_case : Tuple = -10 _snake_case : str = self.image_processor.size["longest_edge"] def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_ = None , **lowercase_ , ): _snake_case : List[Any] = self.image_processor( lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # pop arguments that are not used in the foward but used nevertheless _snake_case : Any = encoding_image_processor["original_sizes"] if hasattr(lowercase_ , "numpy" ): # Checks if Torch or TF tensor _snake_case : int = original_sizes.numpy() _snake_case ,_snake_case ,_snake_case : Union[str, Any] = self._check_and_preprocess_points( input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , ) _snake_case : Dict = self._normalize_and_convert( lowercase_ , lowercase_ , input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , return_tensors=lowercase_ , ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_="pt" , ): if input_points is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : int = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] ) for point in input_points ] else: _snake_case : Dict = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ ) for point, original_size in zip(lowercase_ , lowercase_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _snake_case ,_snake_case : int = self._pad_points_and_labels(lowercase_ , lowercase_ ) _snake_case : Any = np.array(lowercase_ ) if input_labels is not None: _snake_case : Optional[Any] = np.array(lowercase_ ) if input_boxes is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : Optional[Any] = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] , is_bounding_box=lowercase_ ) for box in input_boxes ] else: _snake_case : List[str] = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ , is_bounding_box=lowercase_ ) for box, original_size in zip(lowercase_ , lowercase_ ) ] _snake_case : Tuple = np.array(lowercase_ ) if input_boxes is not None: if return_tensors == "pt": _snake_case : List[str] = torch.from_numpy(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[int] = tf.expand_dims(lowercase_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": _snake_case : Tuple = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : int = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : Tuple = tf.expand_dims(lowercase_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": _snake_case : Dict = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : str = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _snake_case : Optional[Any] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : List[Any] = tf.expand_dims(lowercase_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : List[Any] = max([point.shape[0] for point in input_points] ) _snake_case : List[str] = [] for i, point in enumerate(lowercase_ ): if point.shape[0] != expected_nb_points: _snake_case : Optional[Any] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _snake_case : Union[str, Any] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowercase_ ) _snake_case : Optional[Any] = processed_input_points return input_points, input_labels def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=False ): _snake_case ,_snake_case : Optional[int] = original_size _snake_case ,_snake_case : List[str] = self.image_processor._get_preprocess_shape(lowercase_ , longest_edge=lowercase_ ) _snake_case : Optional[Any] = deepcopy(lowercase_ ).astype(lowercase_ ) if is_bounding_box: _snake_case : str = coords.reshape(-1 , 2 , 2 ) _snake_case : Optional[Any] = coords[..., 0] * (new_w / old_w) _snake_case : Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: _snake_case : Optional[Any] = coords.reshape(-1 , 4 ) return coords def UpperCamelCase ( self , lowercase_=None , lowercase_=None , lowercase_=None , ): if input_points is not None: if hasattr(lowercase_ , "numpy" ): # Checks for TF or Torch tensor _snake_case : Union[str, Any] = input_points.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_points[0] , lowercase_ ): raise ValueError("Input points must be a list of list of floating points." ) _snake_case : Any = [np.array(lowercase_ ) for input_point in input_points] else: _snake_case : Optional[int] = None if input_labels is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : Tuple = input_labels.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_labels[0] , lowercase_ ): raise ValueError("Input labels must be a list of list integers." ) _snake_case : Tuple = [np.array(lowercase_ ) for label in input_labels] else: _snake_case : Optional[Any] = None if input_boxes is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : List[str] = input_boxes.numpy().tolist() if ( not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_boxes[0] , lowercase_ ) or not isinstance(input_boxes[0][0] , lowercase_ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) _snake_case : List[Any] = [np.array(lowercase_ ).astype(np.floataa ) for box in input_boxes] else: _snake_case : Optional[int] = None return input_points, input_labels, input_boxes @property def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(lowercase_ ) ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.image_processor.post_process_masks(*lowercase_ , **lowercase_ )
670
1
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = BioGptTokenizer _lowerCamelCase = False def UpperCamelCase ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _snake_case : int = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] _snake_case : List[str] = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Optional[int] = ["l o 123", "lo w 1456", "e r</w> 1789", ""] _snake_case : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" ) as fp: fp.write(json.dumps(lowercase_ ) ) with open(self.merges_file , "w" ) as fp: fp.write("\n".join(lowercase_ ) ) def UpperCamelCase ( self , lowercase_ ): _snake_case : int = "lower newer" _snake_case : Dict = "lower newer" return input_text, output_text def UpperCamelCase ( self ): _snake_case : Union[str, Any] = BioGptTokenizer(self.vocab_file , self.merges_file ) _snake_case : Union[str, Any] = "lower" _snake_case : Optional[Any] = ["low", "er</w>"] _snake_case : List[str] = tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Optional[Any] = tokens + ["<unk>"] _snake_case : Tuple = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase_ ) , lowercase_ ) @slow def UpperCamelCase ( self ): _snake_case : Tuple = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) _snake_case : str = tokenizer.encode("sequence builders" , add_special_tokens=lowercase_ ) _snake_case : Any = tokenizer.encode("multi-sequence build" , add_special_tokens=lowercase_ ) _snake_case : Any = tokenizer.build_inputs_with_special_tokens(lowercase_ ) _snake_case : Optional[int] = tokenizer.build_inputs_with_special_tokens(lowercase_ , lowercase_ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
670
def snake_case (__lowercase ) -> int: '''simple docstring''' if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _snake_case : Union[str, Any] = grid[0] for row_n in range(1 , len(__lowercase ) ): _snake_case : Union[str, Any] = grid[row_n] _snake_case : List[Any] = fill_row(__lowercase , __lowercase ) _snake_case : List[Any] = grid[row_n] return grid[-1][-1] def snake_case (__lowercase , __lowercase ) -> list: '''simple docstring''' current_row[0] += row_above[0] for cell_n in range(1 , len(__lowercase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
670
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class lowercase_ : _lowerCamelCase = MBartConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , ): _snake_case : str = parent _snake_case : Optional[Any] = batch_size _snake_case : int = seq_length _snake_case : List[Any] = is_training _snake_case : Union[str, Any] = use_labels _snake_case : Union[str, Any] = vocab_size _snake_case : str = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : Optional[Any] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_dropout_prob _snake_case : Dict = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : List[Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Union[str, Any] = bos_token_id def UpperCamelCase ( self ): _snake_case : List[str] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : int = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _snake_case : Optional[int] = prepare_mbart_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Any = TFMBartModel(config=lowercase_ ).get_decoder() _snake_case : Tuple = inputs_dict["input_ids"] _snake_case : List[str] = input_ids[:1, :] _snake_case : Optional[int] = inputs_dict["attention_mask"][:1, :] _snake_case : Any = inputs_dict["head_mask"] _snake_case : Any = 1 # first forward pass _snake_case : Union[str, Any] = model(lowercase_ , attention_mask=lowercase_ , head_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : List[str] = outputs.to_tuple() _snake_case : Tuple = past_key_values[1] def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> Tuple: '''simple docstring''' if attention_mask is None: _snake_case : Optional[int] = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case : Union[str, Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () _lowerCamelCase = (TFMBartForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFMBartModelTester(self ) _snake_case : Union[str, Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) @require_sentencepiece @require_tokenizers @require_tf class lowercase_ ( unittest.TestCase ): _lowerCamelCase = [ ' UN Chief Says There Is No Military Solution in Syria', ] _lowerCamelCase = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] _lowerCamelCase = 'facebook/mbart-large-en-ro' @cached_property def UpperCamelCase ( self ): return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCamelCase ( self ): _snake_case : Union[str, Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def UpperCamelCase ( self , **lowercase_ ): _snake_case : int = self.translate_src_text(**lowercase_ ) self.assertListEqual(self.expected_text , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): _snake_case : Union[str, Any] = self.tokenizer(self.src_text , **lowercase_ , return_tensors="tf" ) _snake_case : Any = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) _snake_case : str = self.tokenizer.batch_decode(lowercase_ , skip_special_tokens=lowercase_ ) return generated_words @slow def UpperCamelCase ( self ): self._assert_generated_batch_equal_expected()
670
import random def snake_case (__lowercase , __lowercase ) -> tuple: '''simple docstring''' _snake_case ,_snake_case ,_snake_case : List[Any] = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def snake_case (__lowercase , __lowercase ) -> List[Any]: '''simple docstring''' if index >= len(__lowercase ) or index < 0: return None _snake_case : Any = items[random.randint(0 , len(__lowercase ) - 1 )] _snake_case : Tuple = 0 _snake_case ,_snake_case ,_snake_case : Tuple = _partition(__lowercase , __lowercase ) _snake_case : Tuple = len(__lowercase ) _snake_case : List[str] = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
670
1
from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def snake_case (__lowercase , __lowercase , __lowercase = False ) -> list[float]: '''simple docstring''' if radian_mode: return [magnitude * cos(__lowercase ), magnitude * sin(__lowercase )] return [magnitude * cos(radians(__lowercase ) ), magnitude * sin(radians(__lowercase ) )] def snake_case (__lowercase , __lowercase , __lowercase = 10**-1 ) -> bool: '''simple docstring''' _snake_case : NDArray[floataa] = cross(__lowercase , __lowercase ) _snake_case : float = sum(__lowercase ) return abs(__lowercase ) < eps if __name__ == "__main__": # Test to check if it works __SCREAMING_SNAKE_CASE : List[str] = array( [ polar_force(7_18.4, 1_8_0 - 3_0), polar_force(8_79.54, 4_5), polar_force(1_0_0, -9_0), ] ) __SCREAMING_SNAKE_CASE : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg __SCREAMING_SNAKE_CASE : Tuple = array( [ polar_force(3_0 * 9.81, 1_5), polar_force(2_1_5, 1_8_0 - 4_5), polar_force(2_6_4, 9_0 - 3_0), ] ) __SCREAMING_SNAKE_CASE : Optional[int] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg __SCREAMING_SNAKE_CASE : Optional[Any] = array([[0, -2_0_0_0], [0, -1_2_0_0], [0, 1_5_6_0_0], [0, -1_2_4_0_0]]) __SCREAMING_SNAKE_CASE : str = array([[0, 0], [6, 0], [1_0, 0], [1_2, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
670
from math import pow, sqrt def snake_case (*__lowercase ) -> bool: '''simple docstring''' _snake_case : str = len(__lowercase ) > 0 and all(value > 0.0 for value in values ) return result def snake_case (__lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
670
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[str] = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Optional[Any] = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
1
import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) __SCREAMING_SNAKE_CASE : int = logging.getLogger() def snake_case () -> int: '''simple docstring''' _snake_case : str = argparse.ArgumentParser() parser.add_argument("-f" ) _snake_case : Optional[int] = parser.parse_args() return args.f def snake_case (__lowercase ) -> List[Any]: '''simple docstring''' _snake_case : Optional[int] = {} _snake_case : Union[str, Any] = os.path.join(__lowercase , "all_results.json" ) if os.path.exists(__lowercase ): with open(__lowercase , "r" ) as f: _snake_case : Union[str, Any] = json.load(__lowercase ) else: raise ValueError(F"""can't find {path}""" ) return results def snake_case () -> Optional[int]: '''simple docstring''' _snake_case : List[Any] = torch.cuda.is_available() and torch_device == "cuda" return is_using_cuda and is_apex_available() __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowercase_ ( __snake_case ): @classmethod def UpperCamelCase ( cls ): # Write Accelerate config, will pick up on CPU, GPU, and multi-GPU _snake_case : Optional[Any] = tempfile.mkdtemp() _snake_case : List[Any] = os.path.join(cls.tmpdir , "default_config.yml" ) write_basic_config(save_location=cls.configPath ) _snake_case : int = ["accelerate", "launch", "--config_file", cls.configPath] @classmethod def UpperCamelCase ( cls ): shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : str = self.get_auto_remove_tmp_dir() _snake_case : Dict = f""" {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --seed=42 --checkpointing_steps epoch --with_tracking """.split() if is_cuda_and_apex_available(): testargs.append("--fp16" ) run_command(self._launch_args + testargs ) _snake_case : Optional[int] = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "glue_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Tuple = self.get_auto_remove_tmp_dir() _snake_case : Union[str, Any] = f""" {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --block_size 128 --per_device_train_batch_size 5 --per_device_eval_batch_size 5 --num_train_epochs 2 --output_dir {tmp_dir} --checkpointing_steps epoch --with_tracking """.split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) _snake_case : List[str] = get_results(lowercase_ ) self.assertLess(result["perplexity"] , 100 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "clm_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : str = self.get_auto_remove_tmp_dir() _snake_case : List[str] = f""" {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --num_train_epochs=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : List[Any] = get_results(lowercase_ ) self.assertLess(result["perplexity"] , 42 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "mlm_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu _snake_case : Optional[Any] = 7 if get_gpu_count() > 1 else 2 _snake_case : Tuple = self.get_auto_remove_tmp_dir() _snake_case : Any = f""" {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : List[Any] = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) self.assertLess(result["train_loss"] , 0.5 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "ner_no_trainer" ) ) ) @unittest.skip(reason="Fix me @muellerzr" ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Tuple = self.get_auto_remove_tmp_dir() _snake_case : str = f""" {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --seed=42 --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : Tuple = get_results(lowercase_ ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result["eval_f1"] , 28 ) self.assertGreaterEqual(result["eval_exact"] , 28 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "qa_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_auto_remove_tmp_dir() _snake_case : Optional[int] = f""" {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/swag/sample.json --validation_file tests/fixtures/tests_samples/swag/sample.json --output_dir {tmp_dir} --max_train_steps=20 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : Tuple = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_accuracy"] , 0.8 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "swag_no_trainer" ) ) ) @slow @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_auto_remove_tmp_dir() _snake_case : Dict = f""" {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : Any = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_rouge1"] , 10 ) self.assertGreaterEqual(result["eval_rouge2"] , 2 ) self.assertGreaterEqual(result["eval_rougeL"] , 7 ) self.assertGreaterEqual(result["eval_rougeLsum"] , 7 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "summarization_no_trainer" ) ) ) @slow @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Tuple = self.get_auto_remove_tmp_dir() _snake_case : Any = f""" {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py --model_name_or_path sshleifer/student_marian_en_ro_6_1 --source_lang en --target_lang ro --train_file tests/fixtures/tests_samples/wmt16/sample.json --validation_file tests/fixtures/tests_samples/wmt16/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --num_beams=6 --learning_rate=3e-3 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --source_lang en_XX --target_lang ro_RO --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) _snake_case : str = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_bleu"] , 30 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "translation_no_trainer" ) ) ) @slow def UpperCamelCase ( self ): _snake_case : Union[str, Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase_ ) _snake_case : Tuple = self.get_auto_remove_tmp_dir() _snake_case : Tuple = f""" {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py --dataset_name huggingface/semantic-segmentation-test-sample --output_dir {tmp_dir} --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch """.split() run_command(self._launch_args + testargs ) _snake_case : str = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_overall_accuracy"] , 0.10 ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_auto_remove_tmp_dir() _snake_case : Tuple = f""" {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py --model_name_or_path google/vit-base-patch16-224-in21k --dataset_name hf-internal-testing/cats_vs_dogs_sample --learning_rate 1e-4 --per_device_train_batch_size 2 --per_device_eval_batch_size 1 --max_train_steps 2 --train_val_split 0.1 --seed 42 --output_dir {tmp_dir} --with_tracking --checkpointing_steps 1 """.split() if is_cuda_and_apex_available(): testargs.append("--fp16" ) run_command(self._launch_args + testargs ) _snake_case : Optional[int] = get_results(lowercase_ ) # The base model scores a 25% self.assertGreaterEqual(result["eval_accuracy"] , 0.6 ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "step_1" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase_ , "image_classification_no_trainer" ) ) )
670
from __future__ import annotations from typing import TypedDict class lowercase_ ( __snake_case ): _lowerCamelCase = 42 _lowerCamelCase = 42 def snake_case (__lowercase ) -> list[str]: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(__lowercase ) )] def snake_case (__lowercase ) -> BWTTransformDict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) _snake_case : List[str] = all_rotations(__lowercase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation _snake_case : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__lowercase ), } return response def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: _snake_case : Union[str, Any] = int(__lowercase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(__lowercase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) _snake_case : Optional[Any] = [""] * len(__lowercase ) for _ in range(len(__lowercase ) ): for i in range(len(__lowercase ) ): _snake_case : Tuple = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = 'Provide a string that I will generate its BWT transform: ' __SCREAMING_SNAKE_CASE : Optional[Any] = input(entry_msg).strip() __SCREAMING_SNAKE_CASE : int = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result['bwt_string']}\'''' ) __SCREAMING_SNAKE_CASE : List[str] = reverse_bwt(result['bwt_string'], result['idx_original_string']) print( F'''Reversing Burrows Wheeler transform for entry \'{result['bwt_string']}\' ''' F'''we get original string \'{original_string}\'''' )
670
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Union[str, Any] = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class lowercase_ ( __snake_case ): _lowerCamelCase = 'switch_transformers' _lowerCamelCase = ['past_key_values'] _lowerCamelCase = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , lowercase_=32_128 , lowercase_=768 , lowercase_=64 , lowercase_=2_048 , lowercase_=64 , lowercase_=12 , lowercase_=3 , lowercase_=12 , lowercase_=3 , lowercase_=12 , lowercase_=8 , lowercase_=False , lowercase_=0.01 , lowercase_="float32" , lowercase_=False , lowercase_=32 , lowercase_=128 , lowercase_=0.1 , lowercase_=1e-6 , lowercase_=0.001 , lowercase_=0.001 , lowercase_=1.0 , lowercase_="relu" , lowercase_=True , lowercase_=False , lowercase_=True , lowercase_=0 , lowercase_=1 , **lowercase_ , ): _snake_case : Optional[Any] = vocab_size _snake_case : Dict = d_model _snake_case : Any = d_kv _snake_case : Union[str, Any] = d_ff _snake_case : int = num_sparse_encoder_layers _snake_case : int = num_layers _snake_case : List[Any] = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _snake_case : str = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: _snake_case : int = self.num_layers // self.num_sparse_encoder_layers else: _snake_case : int = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: _snake_case : Optional[int] = self.num_decoder_layers // self.num_sparse_decoder_layers else: _snake_case : Tuple = self.num_decoder_layers # HACK: this will create 0 sparse layers _snake_case : Tuple = num_heads _snake_case : List[str] = num_experts _snake_case : List[Any] = expert_capacity _snake_case : Any = router_bias _snake_case : List[str] = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) _snake_case : Optional[int] = router_dtype _snake_case : Any = router_ignore_padding_tokens _snake_case : Tuple = relative_attention_num_buckets _snake_case : Dict = relative_attention_max_distance _snake_case : int = dropout_rate _snake_case : str = layer_norm_epsilon _snake_case : Any = initializer_factor _snake_case : Union[str, Any] = feed_forward_proj _snake_case : Dict = use_cache _snake_case : Union[str, Any] = add_router_probs _snake_case : int = router_z_loss_coef _snake_case : int = router_aux_loss_coef _snake_case : List[Any] = self.feed_forward_proj.split("-" ) _snake_case : Dict = act_info[-1] _snake_case : Optional[int] = act_info[0] == "gated" if len(lowercase_ ) > 1 and act_info[0] != "gated" or len(lowercase_ ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "'gated-gelu' or 'relu'" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": _snake_case : List[Any] = "gelu_new" super().__init__( pad_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , **lowercase_ , )
670
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( 'stable diffusion controlnet', '0.22.0', 'Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.', standard_warn=False, stacklevel=3, )
670
1
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def snake_case (__lowercase ) -> tuple: '''simple docstring''' return (data["data"], data["target"]) def snake_case (__lowercase , __lowercase , __lowercase ) -> np.ndarray: '''simple docstring''' _snake_case : Dict = XGBRegressor(verbosity=0 , random_state=42 ) xgb.fit(__lowercase , __lowercase ) # Predict target for test data _snake_case : Tuple = xgb.predict(__lowercase ) _snake_case : Optional[Any] = predictions.reshape(len(__lowercase ) , 1 ) return predictions def snake_case () -> None: '''simple docstring''' _snake_case : List[str] = fetch_california_housing() _snake_case ,_snake_case : int = data_handling(__lowercase ) _snake_case ,_snake_case ,_snake_case ,_snake_case : Tuple = train_test_split( __lowercase , __lowercase , test_size=0.25 , random_state=1 ) _snake_case : Union[str, Any] = xgboost(__lowercase , __lowercase , __lowercase ) # Error printing print(F"""Mean Absolute Error : {mean_absolute_error(__lowercase , __lowercase )}""" ) print(F"""Mean Square Error : {mean_squared_error(__lowercase , __lowercase )}""" ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
670
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase_ : _lowerCamelCase = LEDConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=4 , ): _snake_case : Optional[int] = parent _snake_case : str = batch_size _snake_case : int = seq_length _snake_case : Dict = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = vocab_size _snake_case : str = hidden_size _snake_case : int = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : int = intermediate_size _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : int = max_position_embeddings _snake_case : Union[str, Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Any = bos_token_id _snake_case : str = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after _snake_case : List[Any] = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests _snake_case : List[str] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def UpperCamelCase ( self ): _snake_case : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Any = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) _snake_case : Optional[Any] = prepare_led_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) _snake_case : int = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]] , axis=-1 , ) _snake_case : List[Any] = global_attention_mask return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = TFLEDModel(config=lowercase_ ).get_decoder() _snake_case : Optional[Any] = inputs_dict["input_ids"] _snake_case : Optional[int] = input_ids[:1, :] _snake_case : int = inputs_dict["attention_mask"][:1, :] _snake_case : int = 1 # first forward pass _snake_case : str = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case : Tuple = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case : List[str] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case : str = model(lowercase_ , attention_mask=lowercase_ )[0] _snake_case : List[str] = model(lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_ , lowercase_ , rtol=1e-3 ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> List[Any]: '''simple docstring''' if attention_mask is None: _snake_case : int = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () _lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFLEDModelTester(self ) _snake_case : List[Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = tf.zeros_like(inputs_dict["attention_mask"] ) _snake_case : Tuple = 2 _snake_case : Dict = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) _snake_case : Tuple = True _snake_case : Union[str, Any] = self.model_tester.seq_length _snake_case : Union[str, Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ ): _snake_case : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(lowercase_ ): _snake_case : int = [t.numpy() for t in outputs.encoder_attentions] _snake_case : Optional[int] = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: _snake_case : Union[str, Any] = True _snake_case : Dict = False _snake_case : Any = False _snake_case : Any = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) _snake_case : Tuple = len(lowercase_ ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: _snake_case : int = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Any = model_class(lowercase_ ) _snake_case : Optional[Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine _snake_case : Optional[int] = True _snake_case : Optional[int] = True _snake_case : List[Any] = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # TODO: Head-masking not yet implement pass def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' return tf.constant(__lowercase , dtype=tf.intaa ) __SCREAMING_SNAKE_CASE : List[Any] = 1E-4 @slow @require_tf class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Dict = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here _snake_case : Union[str, Any] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[int] = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Union[str, Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Optional[Any] = model(**lowercase_ )[0] _snake_case : str = (1, 1_024, 768) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[Any] = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 ) def UpperCamelCase ( self ): _snake_case : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here _snake_case : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : int = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Tuple = model(**lowercase_ )[0] _snake_case : Any = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[int] = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 , rtol=1e-3 )
670
1
from typing import TYPE_CHECKING from ...utils import _LazyModule __SCREAMING_SNAKE_CASE : List[Any] = {'tokenization_byt5': ['ByT5Tokenizer']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
1
def snake_case (__lowercase ) -> list: '''simple docstring''' def merge(__lowercase , __lowercase ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(__lowercase ) <= 1: return collection _snake_case : Union[str, Any] = len(__lowercase ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : int = input('Enter numbers separated by a comma:\n').strip() __SCREAMING_SNAKE_CASE : Any = [int(item) for item in user_input.split(',')] print(*merge_sort(unsorted), sep=',')
670
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Any = tempfile.mkdtemp() # fmt: off _snake_case : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _snake_case : Dict = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] _snake_case : Optional[int] = {"unk_token": "<unk>"} _snake_case : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) _snake_case : Any = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } _snake_case : Optional[Any] = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Union[str, Any] = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Tuple = self.get_tokenizer() _snake_case : Any = self.get_rust_tokenizer() _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) _snake_case : Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) _snake_case : List[Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : List[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Tuple = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : int = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Optional[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : str = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : Dict = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : int = processor(text=lowercase_ ) _snake_case : str = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[Any] = "lower newer" _snake_case : int = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[str] = self.get_tokenizer() _snake_case : Union[str, Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Dict = self.prepare_image_inputs() _snake_case : List[Any] = processor(images=lowercase_ , visual_prompt=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[Any] = self.get_tokenizer() _snake_case : str = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Any = processor.batch_decode(lowercase_ ) _snake_case : Any = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ )
670
1
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def snake_case (__lowercase , __lowercase ) -> Union[str, Any]: '''simple docstring''' _snake_case : Optional[Any] = torch.load(__lowercase , map_location="cpu" ) _snake_case : List[Any] = chkpt["model"] # We have the base model one level deeper than the original XLM repository _snake_case : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: _snake_case : List[str] = v else: _snake_case : Union[str, Any] = v _snake_case : Dict = chkpt["params"] _snake_case : Tuple = {n: v for n, v in config.items() if not isinstance(__lowercase , (torch.FloatTensor, numpy.ndarray) )} _snake_case : int = chkpt["dico_word2id"] _snake_case : Union[str, Any] = {s + "</w>" if s.find("@@" ) == -1 and i > 13 else s.replace("@@" , "" ): i for s, i in vocab.items()} # Save pytorch-model _snake_case : Union[str, Any] = pytorch_dump_folder_path + "/" + WEIGHTS_NAME _snake_case : Dict = pytorch_dump_folder_path + "/" + CONFIG_NAME _snake_case : Tuple = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(__lowercase , __lowercase ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(__lowercase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowercase , indent=2 ) + "\n" ) print(F"""Save vocab file to {pytorch_config_dump_path}""" ) with open(__lowercase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowercase , indent=2 ) + "\n" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
670
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
1
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ ): super().__init__() _snake_case : List[str] = nn.ModuleList(lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = True , ): for i, (image, scale, controlnet) in enumerate(zip(lowercase_ , lowercase_ , self.nets ) ): _snake_case ,_snake_case : Optional[int] = controlnet( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) # merge samples if i == 0: _snake_case ,_snake_case : Tuple = down_samples, mid_sample else: _snake_case : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase_ , lowercase_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , ): _snake_case : Tuple = 0 _snake_case : Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase_ , is_main_process=lowercase_ , save_function=lowercase_ , safe_serialization=lowercase_ , variant=lowercase_ , ) idx += 1 _snake_case : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : List[str] = 0 _snake_case : Optional[Any] = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case : Optional[Any] = pretrained_model_path while os.path.isdir(lowercase_ ): _snake_case : int = ControlNetModel.from_pretrained(lowercase_ , **lowercase_ ) controlnets.append(lowercase_ ) idx += 1 _snake_case : str = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(lowercase_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(lowercase_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(lowercase_ )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(lowercase_ )
670
1
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def snake_case () -> Optional[Any]: '''simple docstring''' _snake_case : int = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" _snake_case : List[Any] = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ).convert("RGB" ) return image def snake_case (__lowercase ) -> int: '''simple docstring''' _snake_case : Dict = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.weight""", F"""vision_model.encoder.layers.{i}.layer_norm1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.bias""", F"""vision_model.encoder.layers.{i}.layer_norm1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.weight""", F"""vision_model.encoder.layers.{i}.layer_norm2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.bias""", F"""vision_model.encoder.layers.{i}.layer_norm2.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.qkv.weight""", F"""vision_model.encoder.layers.{i}.self_attn.qkv.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.weight""", F"""vision_model.encoder.layers.{i}.self_attn.projection.weight""",) ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.bias""", F"""vision_model.encoder.layers.{i}.self_attn.projection.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc2.bias""") ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def snake_case (__lowercase , __lowercase , __lowercase ) -> List[Any]: '''simple docstring''' _snake_case : int = dct.pop(__lowercase ) _snake_case : Optional[int] = val def snake_case (__lowercase , __lowercase ) -> Tuple: '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases _snake_case : Optional[int] = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.q_bias""" ) _snake_case : int = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.v_bias""" ) # next, set bias in the state dict _snake_case : int = torch.cat((q_bias, torch.zeros_like(__lowercase , requires_grad=__lowercase ), v_bias) ) _snake_case : int = qkv_bias def snake_case (__lowercase , __lowercase ) -> List[str]: '''simple docstring''' _snake_case : int = 364 if "coco" in model_name else 224 _snake_case : List[Any] = BlipaVisionConfig(image_size=__lowercase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: _snake_case : Optional[int] = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=__lowercase ).to_dict() elif "opt-6.7b" in model_name: _snake_case : Optional[int] = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=__lowercase ).to_dict() elif "t5-xl" in model_name: _snake_case : Optional[int] = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: _snake_case : Tuple = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() _snake_case : str = BlipaConfig(vision_config=__lowercase , text_config=__lowercase ) return config, image_size @torch.no_grad() def snake_case (__lowercase , __lowercase=None , __lowercase=False ) -> Dict: '''simple docstring''' _snake_case : Tuple = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) _snake_case : Tuple = tokenizer("\n" , add_special_tokens=__lowercase ).input_ids[0] _snake_case ,_snake_case : List[str] = get_blipa_config(__lowercase , eos_token_id=__lowercase ) _snake_case : Dict = BlipaForConditionalGeneration(__lowercase ).eval() _snake_case : Optional[Any] = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } _snake_case ,_snake_case : Union[str, Any] = model_name_to_original[model_name] # load original model print("Loading original model..." ) _snake_case : List[str] = "cuda" if torch.cuda.is_available() else "cpu" _snake_case ,_snake_case ,_snake_case : Tuple = load_model_and_preprocess( name=__lowercase , model_type=__lowercase , is_eval=__lowercase , device=__lowercase ) original_model.eval() print("Done!" ) # update state dict keys _snake_case : str = original_model.state_dict() _snake_case : Tuple = create_rename_keys(__lowercase ) for src, dest in rename_keys: rename_key(__lowercase , __lowercase , __lowercase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): _snake_case : str = state_dict.pop(__lowercase ) if key.startswith("Qformer.bert" ): _snake_case : Any = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: _snake_case : Optional[int] = key.replace("self" , "attention" ) if "opt_proj" in key: _snake_case : Tuple = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: _snake_case : str = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): _snake_case : Union[str, Any] = key.replace("opt" , "language" ) if key.startswith("t5" ): _snake_case : List[str] = key.replace("t5" , "language" ) _snake_case : Tuple = val # read in qv biases read_in_q_v_bias(__lowercase , __lowercase ) _snake_case ,_snake_case : int = hf_model.load_state_dict(__lowercase , strict=__lowercase ) assert len(__lowercase ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] _snake_case : List[Any] = load_demo_image() _snake_case : Dict = vis_processors["eval"](__lowercase ).unsqueeze(0 ).to(__lowercase ) _snake_case : str = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(__lowercase ) # create processor _snake_case : Optional[Any] = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=__lowercase , image_std=__lowercase ) _snake_case : str = BlipaProcessor(image_processor=__lowercase , tokenizer=__lowercase ) _snake_case : str = processor(images=__lowercase , return_tensors="pt" ).pixel_values.to(__lowercase ) # make sure processor creates exact same pixel values assert torch.allclose(__lowercase , __lowercase ) original_model.to(__lowercase ) hf_model.to(__lowercase ) with torch.no_grad(): if "opt" in model_name: _snake_case : Optional[Any] = original_model({"image": original_pixel_values, "text_input": [""]} ).logits _snake_case : Optional[int] = hf_model(__lowercase , __lowercase ).logits else: _snake_case : Optional[int] = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits _snake_case : Optional[int] = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) _snake_case : int = hf_model(__lowercase , __lowercase , labels=__lowercase ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": _snake_case : Optional[Any] = torch.tensor( [[-41.5850, -4.4440, -8.9922], [-47.4322, -5.9143, -1.7340]] , device=__lowercase ) assert torch.allclose(logits[0, :3, :3] , __lowercase , atol=1e-4 ) elif model_name == "blip2-flan-t5-xl-coco": _snake_case : List[str] = torch.tensor( [[-57.0109, -9.8967, -12.6280], [-68.6578, -12.7191, -10.5065]] , device=__lowercase ) else: # cast to same type _snake_case : List[Any] = logits.dtype assert torch.allclose(original_logits.to(__lowercase ) , __lowercase , atol=1e-2 ) print("Looks ok!" ) print("Generating a caption..." ) _snake_case : str = "" _snake_case : str = tokenizer(__lowercase , return_tensors="pt" ).input_ids.to(__lowercase ) _snake_case : str = original_model.generate({"image": original_pixel_values} ) _snake_case : Dict = hf_model.generate( __lowercase , __lowercase , do_sample=__lowercase , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , __lowercase ) _snake_case : Optional[int] = input_ids.shape[1] _snake_case : Optional[Any] = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=__lowercase ) _snake_case : List[str] = [text.strip() for text in output_text] print("HF generation:" , __lowercase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(__lowercase ) hf_model.save_pretrained(__lowercase ) if push_to_hub: processor.push_to_hub(F"""nielsr/{model_name}""" ) hf_model.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() __SCREAMING_SNAKE_CASE : Union[str, Any] = [ 'blip2-opt-2.7b', 'blip2-opt-6.7b', 'blip2-opt-2.7b-coco', 'blip2-opt-6.7b-coco', 'blip2-flan-t5-xl', 'blip2-flan-t5-xl-coco', 'blip2-flan-t5-xxl', ] parser.add_argument( '--model_name', default='blip2-opt-2.7b', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) __SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
670
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'CLIPImageProcessor' _lowerCamelCase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Dict = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _snake_case : str = self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if images is not None: _snake_case : List[str] = self.image_processor(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if text is not None and images is not None: _snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase_ ) , tensor_type=lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
1
import random def snake_case (__lowercase , __lowercase , __lowercase = False ) -> dict: '''simple docstring''' _snake_case : dict = {i: [] for i in range(__lowercase )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(__lowercase ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(__lowercase ): for j in range(i + 1 , __lowercase ): if random.random() < probability: graph[i].append(__lowercase ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(__lowercase ) return graph def snake_case (__lowercase ) -> dict: '''simple docstring''' return { i: [j for j in range(__lowercase ) if i != j] for i in range(__lowercase ) } if __name__ == "__main__": import doctest doctest.testmod()
670
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
670
1
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = BarthezTokenizer _lowerCamelCase = BarthezTokenizerFast _lowerCamelCase = True _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Tuple = BarthezTokenizerFast.from_pretrained("moussaKam/mbarthez" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=lowercase_ ) _snake_case : int = tokenizer def UpperCamelCase ( self ): _snake_case : List[str] = "<pad>" _snake_case : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(lowercase_ ) , 101_122 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 101_122 ) @require_torch def UpperCamelCase ( self ): _snake_case : Any = ["A long paragraph for summarization.", "Another paragraph for summarization."] _snake_case : Optional[int] = [0, 57, 3_018, 70_307, 91, 2] _snake_case : Optional[int] = self.tokenizer( lowercase_ , max_length=len(lowercase_ ) , padding=lowercase_ , truncation=lowercase_ , return_tensors="pt" ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) _snake_case : str = batch.input_ids.tolist()[0] self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : Dict = self.get_rust_tokenizer() _snake_case : Union[str, Any] = "I was born in 92000, and this is falsé." _snake_case : str = tokenizer.tokenize(lowercase_ ) _snake_case : Optional[int] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Dict = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Optional[Any] = self.get_rust_tokenizer() _snake_case : int = tokenizer.encode(lowercase_ ) _snake_case : Dict = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Tuple = {"input_ids": [[0, 490, 14_328, 4_507, 354, 47, 43_669, 95, 25, 78_117, 20_215, 19_779, 190, 22, 400, 4, 35_343, 80_310, 603, 86, 24_937, 105, 33_438, 94_762, 196, 39_642, 7, 15, 15_933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10_534, 87, 25, 66, 3_358, 196, 55_289, 8, 82_961, 81, 2_204, 75_203, 7, 15, 763, 12_956, 216, 178, 14_328, 9_595, 1_377, 69_693, 7, 448, 71_021, 196, 18_106, 1_437, 13_974, 108, 9_083, 4, 49_315, 7, 39, 86, 1_326, 2_793, 46_333, 4, 448, 196, 74_588, 7, 49_315, 7, 39, 21, 822, 38_470, 74, 21, 66_723, 62_480, 8, 22_050, 5, 2]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. _snake_case : str = [ "Le transformeur est un modèle d'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="moussaKam/mbarthez" , revision="c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6" , sequences=lowercase_ , )
670
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def snake_case (*__lowercase ) -> Dict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): _snake_case : Dict = list(__lowercase ) for i in range(len(__lowercase ) ): _snake_case : List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : str = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowercase , __lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def snake_case (__lowercase = None , __lowercase = 128 ) -> Any: '''simple docstring''' if function is None: return functools.partial(__lowercase , starting_batch_size=__lowercase ) _snake_case : List[str] = starting_batch_size def decorator(*__lowercase , **__lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _snake_case : Optional[Any] = list(inspect.signature(__lowercase ).parameters.keys() ) # Guard against user error if len(__lowercase ) < (len(__lowercase ) + 1): _snake_case : str = ", ".join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowercase , *__lowercase , **__lowercase ) except Exception as e: if should_reduce_batch_size(__lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
670
1
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Tuple = inspect.getfile(accelerate.test_utils ) _snake_case : Optional[int] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) _snake_case : Optional[int] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) _snake_case : Any = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def UpperCamelCase ( self ): print(f"""Found {torch.cuda.device_count()} devices.""" ) _snake_case : Tuple = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase_ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase ( self ): print(f"""Found {torch.cuda.device_count()} devices.""" ) _snake_case : Any = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.operation_file_path] print(f"""Command: {cmd}""" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase_ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase ( self ): _snake_case : str = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(lowercase_ , env=os.environ.copy() ) @require_multi_gpu def UpperCamelCase ( self ): print(f"""Found {torch.cuda.device_count()} devices, using 2 devices only""" ) _snake_case : List[Any] = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(lowercase_ , env=os.environ.copy() ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = Accelerator() __SCREAMING_SNAKE_CASE : Dict = (accelerator.state.process_index + 2, 1_0) __SCREAMING_SNAKE_CASE : Any = torch.randint(0, 1_0, shape).to(accelerator.device) __SCREAMING_SNAKE_CASE : Optional[Any] = '' __SCREAMING_SNAKE_CASE : Dict = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." __SCREAMING_SNAKE_CASE : Any = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." __SCREAMING_SNAKE_CASE : str = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
670
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
1
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __SCREAMING_SNAKE_CASE : Tuple = logging.getLogger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'token-classification' def __init__( self , lowercase_ ): if type(lowercase_ ) == dict: _snake_case : Dict = Namespace(**lowercase_ ) _snake_case : List[str] = import_module("tasks" ) try: _snake_case : Any = getattr(lowercase_ , hparams.task_type ) _snake_case : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) _snake_case : Optional[int] = self.token_classification_task.get_labels(hparams.labels ) _snake_case : List[str] = CrossEntropyLoss().ignore_index super().__init__(lowercase_ , len(self.labels ) , self.mode ) def UpperCamelCase ( self , **lowercase_ ): return self.model(**lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Optional[Any] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": _snake_case : List[str] = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids _snake_case : Tuple = self(**lowercase_ ) _snake_case : Tuple = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def UpperCamelCase ( self ): _snake_case : int = self.hparams for mode in ["train", "dev", "test"]: _snake_case : List[str] = self._feature_file(lowercase_ ) if os.path.exists(lowercase_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , lowercase_ ) _snake_case : Optional[int] = torch.load(lowercase_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) _snake_case : Tuple = self.token_classification_task.read_examples_from_file(args.data_dir , lowercase_ ) _snake_case : List[Any] = self.token_classification_task.convert_examples_to_features( lowercase_ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=lowercase_ , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , lowercase_ ) torch.save(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = False ): _snake_case : Optional[int] = self._feature_file(lowercase_ ) logger.info("Loading features from cached file %s" , lowercase_ ) _snake_case : Tuple = torch.load(lowercase_ ) _snake_case : List[str] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) _snake_case : Dict = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: _snake_case : List[Any] = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: _snake_case : int = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) _snake_case : Union[str, Any] = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) , batch_size=lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): """Compute validation""" "" _snake_case : List[Any] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": _snake_case : List[Any] = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids _snake_case : str = self(**lowercase_ ) _snake_case ,_snake_case : Optional[Any] = outputs[:2] _snake_case : List[str] = logits.detach().cpu().numpy() _snake_case : List[Any] = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase ( self , lowercase_ ): _snake_case : List[str] = torch.stack([x["val_loss"] for x in outputs] ).mean() _snake_case : str = np.concatenate([x["pred"] for x in outputs] , axis=0 ) _snake_case : List[str] = np.argmax(lowercase_ , axis=2 ) _snake_case : List[str] = np.concatenate([x["target"] for x in outputs] , axis=0 ) _snake_case : str = dict(enumerate(self.labels ) ) _snake_case : Optional[Any] = [[] for _ in range(out_label_ids.shape[0] )] _snake_case : Tuple = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) _snake_case : Dict = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(lowercase_ , lowercase_ ), "precision": precision_score(lowercase_ , lowercase_ ), "recall": recall_score(lowercase_ , lowercase_ ), "f1": fa_score(lowercase_ , lowercase_ ), } _snake_case : Union[str, Any] = dict(results.items() ) _snake_case : Dict = results return ret, preds_list, out_label_list def UpperCamelCase ( self , lowercase_ ): # when stable _snake_case ,_snake_case ,_snake_case : str = self._eval_end(lowercase_ ) _snake_case : List[Any] = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase ( self , lowercase_ ): # updating to test_epoch_end instead of deprecated test_end _snake_case ,_snake_case ,_snake_case : Union[str, Any] = self._eval_end(lowercase_ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 _snake_case : Union[str, Any] = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase ( lowercase_ , lowercase_ ): # Add NER specific options BaseTransformer.add_model_specific_args(lowercase_ , lowercase_ ) parser.add_argument( "--task_type" , default="NER" , type=lowercase_ , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=128 , type=lowercase_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=lowercase_ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=lowercase_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __SCREAMING_SNAKE_CASE : Dict = NERTransformer.add_model_specific_args(parser, os.getcwd()) __SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() __SCREAMING_SNAKE_CASE : Tuple = NERTransformer(args) __SCREAMING_SNAKE_CASE : Union[str, Any] = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __SCREAMING_SNAKE_CASE : Dict = sorted(glob.glob(os.path.join(args.output_dir, 'checkpoint-epoch=*.ckpt'), recursive=True)) __SCREAMING_SNAKE_CASE : int = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
670
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
1
def snake_case (__lowercase , __lowercase ) -> int: '''simple docstring''' return int((input_a, input_a).count(0 ) == 0 ) def snake_case () -> None: '''simple docstring''' assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
670
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : Tuple = Rectangle(height=0.5 , width=0.5 ) _snake_case : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : List[str] = [mem.copy() for i in range(6 )] _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("CPU" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : int = [mem.copy() for i in range(4 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = Text("GPU" , font_size=24 ) _snake_case : Optional[int] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = Text("Model" , font_size=24 ) _snake_case : Dict = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case : Union[str, Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : Union[str, Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Union[str, Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Optional[Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _snake_case : List[Any] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : int = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Dict = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) _snake_case : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
670
1
import re def snake_case (__lowercase ) -> list: '''simple docstring''' return [char.split() for char in re.split(r"[^ a-z A-Z 0-9 \s]" , str_ )] def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : int = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def snake_case (__lowercase , __lowercase , __lowercase ) -> str: '''simple docstring''' try: _snake_case : Optional[Any] = split_input(__lowercase ) if upper: _snake_case : Tuple = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: _snake_case : Dict = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def snake_case (__lowercase ) -> str: '''simple docstring''' return to_simple_case(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' try: _snake_case : Optional[int] = to_simple_case(__lowercase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' return to_complex_case(__lowercase , __lowercase , "_" ) def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' return to_complex_case(__lowercase , __lowercase , "-" ) if __name__ == "__main__": __import__('doctest').testmod()
670
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
1
from decimal import Decimal, getcontext from math import ceil, factorial def snake_case (__lowercase ) -> str: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) _snake_case : Tuple = precision _snake_case : Union[str, Any] = ceil(precision / 14 ) _snake_case : Optional[Any] = 426_880 * Decimal(10_005 ).sqrt() _snake_case : Any = 1 _snake_case : List[Any] = 13_591_409 _snake_case : Optional[int] = Decimal(__lowercase ) for k in range(1 , __lowercase ): _snake_case : Dict = factorial(6 * k ) // (factorial(3 * k ) * factorial(__lowercase ) ** 3) linear_term += 545_140_134 exponential_term *= -262_537_412_640_768_000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = 5_0 print(F'''The first {n} digits of pi is: {pi(n)}''')
670
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowercase_ ( __snake_case ): _lowerCamelCase = 'roc_bert' def __init__( self , lowercase_=30_522 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3_072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1e-12 , lowercase_=True , lowercase_=0 , lowercase_="absolute" , lowercase_=None , lowercase_=True , lowercase_=True , lowercase_=768 , lowercase_=910 , lowercase_=512 , lowercase_=24_858 , lowercase_=True , **lowercase_ , ): _snake_case : int = vocab_size _snake_case : Union[str, Any] = max_position_embeddings _snake_case : Union[str, Any] = hidden_size _snake_case : Dict = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : List[Any] = hidden_act _snake_case : Optional[int] = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : List[Any] = type_vocab_size _snake_case : int = layer_norm_eps _snake_case : Optional[Any] = use_cache _snake_case : List[Any] = enable_pronunciation _snake_case : Dict = enable_shape _snake_case : Dict = pronunciation_embed_dim _snake_case : Tuple = pronunciation_vocab_size _snake_case : Tuple = shape_embed_dim _snake_case : List[str] = shape_vocab_size _snake_case : Dict = concat_input _snake_case : int = position_embedding_type _snake_case : int = classifier_dropout super().__init__(pad_token_id=lowercase_ , **lowercase_ )
670
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class lowercase_ ( __snake_case ): _lowerCamelCase = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
670
from cva import destroyAllWindows, imread, imshow, waitKey def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case ,_snake_case : int = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(__lowercase ): for j in range(__lowercase ): _snake_case : Optional[Any] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image __SCREAMING_SNAKE_CASE : Optional[Any] = imread('image_data/lena.jpg', 1) # convert to its negative __SCREAMING_SNAKE_CASE : Tuple = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
670
1
__SCREAMING_SNAKE_CASE : int = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __SCREAMING_SNAKE_CASE : List[str] = frozenset(['prompt', 'negative_prompt']) __SCREAMING_SNAKE_CASE : Tuple = frozenset([]) __SCREAMING_SNAKE_CASE : List[Any] = frozenset(['image']) __SCREAMING_SNAKE_CASE : Any = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = frozenset(['image']) __SCREAMING_SNAKE_CASE : Optional[Any] = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = frozenset(['prompt', 'image', 'negative_prompt']) __SCREAMING_SNAKE_CASE : Optional[Any] = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) __SCREAMING_SNAKE_CASE : Optional[Any] = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) __SCREAMING_SNAKE_CASE : Union[str, Any] = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __SCREAMING_SNAKE_CASE : Tuple = frozenset(['image', 'mask_image']) __SCREAMING_SNAKE_CASE : Any = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) __SCREAMING_SNAKE_CASE : Optional[Any] = frozenset(['example_image', 'image', 'mask_image']) __SCREAMING_SNAKE_CASE : int = frozenset(['class_labels']) __SCREAMING_SNAKE_CASE : Tuple = frozenset(['class_labels']) __SCREAMING_SNAKE_CASE : List[str] = frozenset(['batch_size']) __SCREAMING_SNAKE_CASE : List[str] = frozenset([]) __SCREAMING_SNAKE_CASE : Optional[int] = frozenset(['batch_size']) __SCREAMING_SNAKE_CASE : List[str] = frozenset([]) __SCREAMING_SNAKE_CASE : List[str] = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) __SCREAMING_SNAKE_CASE : Any = frozenset(['prompt', 'negative_prompt']) __SCREAMING_SNAKE_CASE : int = frozenset(['input_tokens']) __SCREAMING_SNAKE_CASE : Union[str, Any] = frozenset(['input_tokens'])
670
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __SCREAMING_SNAKE_CASE : List[str] = Mapping[str, np.ndarray] __SCREAMING_SNAKE_CASE : List[Any] = Mapping[str, Any] # Is a nested dict. __SCREAMING_SNAKE_CASE : List[Any] = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class lowercase_ : _lowerCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _lowerCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _lowerCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _lowerCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _lowerCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _lowerCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _lowerCamelCase = None # Templates used to generate this protein (prediction-only) _lowerCamelCase = None # Chain corresponding to each parent _lowerCamelCase = None def snake_case (__lowercase ) -> Protein: '''simple docstring''' _snake_case : str = r"(\[[A-Z]+\]\n)" _snake_case : List[str] = [tag.strip() for tag in re.split(__lowercase , __lowercase ) if len(__lowercase ) > 0] _snake_case : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("\n" ) for l in tags[1::2]] ) _snake_case : List[str] = ["N", "CA", "C"] _snake_case : Any = None _snake_case : Union[str, Any] = None _snake_case : Optional[int] = None for g in groups: if "[PRIMARY]" == g[0]: _snake_case : Tuple = g[1][0].strip() for i in range(len(__lowercase ) ): if seq[i] not in residue_constants.restypes: _snake_case : Tuple = "X" # FIXME: strings are immutable _snake_case : int = np.array( [residue_constants.restype_order.get(__lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: _snake_case : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(__lowercase , g[1][axis].split() ) ) ) _snake_case : Dict = np.array(__lowercase ) _snake_case : Dict = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: _snake_case : int = np.array(list(map({"-": 0, "+": 1}.get , g[1][0].strip() ) ) ) _snake_case : Any = np.zeros( ( len(__lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : Dict = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__lowercase , atom_mask=__lowercase , aatype=__lowercase , residue_index=np.arange(len(__lowercase ) ) , b_factors=__lowercase , ) def snake_case (__lowercase , __lowercase = 0 ) -> List[str]: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[Any] = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) _snake_case : str = prot.parents _snake_case : str = prot.parents_chain_index if parents is not None and parents_chain_index is not None: _snake_case : int = [p for i, p in zip(__lowercase , __lowercase ) if i == chain_id] if parents is None or len(__lowercase ) == 0: _snake_case : Optional[int] = ["N/A"] pdb_headers.append(F"""PARENT {' '.join(__lowercase )}""" ) return pdb_headers def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[int] = pdb_str.split("\n" ) _snake_case : List[str] = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) _snake_case : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: _snake_case : str = [] if prot.parents_chain_index is not None: _snake_case : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(__lowercase ) , [] ) parent_dict[str(__lowercase )].append(__lowercase ) _snake_case : Any = max([int(__lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): _snake_case : Tuple = parent_dict.get(str(__lowercase ) , ["N/A"] ) parents_per_chain.append(__lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: _snake_case : List[str] = [["N/A"]] def make_parent_line(__lowercase ) -> str: return F"""PARENT {' '.join(__lowercase )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) _snake_case : int = 0 for i, l in enumerate(__lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__lowercase ): _snake_case : Tuple = parents_per_chain[chain_counter] else: _snake_case : str = ["N/A"] out_pdb_lines.append(make_parent_line(__lowercase ) ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = residue_constants.restypes + ["X"] def res_atoa(__lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , "UNK" ) _snake_case : Optional[int] = residue_constants.atom_types _snake_case : List[str] = [] _snake_case : Tuple = prot.atom_mask _snake_case : List[str] = prot.aatype _snake_case : int = prot.atom_positions _snake_case : int = prot.residue_index.astype(np.intaa ) _snake_case : List[Any] = prot.b_factors _snake_case : str = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("Invalid aatypes." ) _snake_case : Union[str, Any] = get_pdb_headers(__lowercase ) if len(__lowercase ) > 0: pdb_lines.extend(__lowercase ) _snake_case : Optional[Any] = aatype.shape[0] _snake_case : str = 1 _snake_case : Tuple = 0 _snake_case : int = string.ascii_uppercase _snake_case : Optional[Any] = None # Add all atom sites. for i in range(__lowercase ): _snake_case : Dict = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue _snake_case : List[Any] = "ATOM" _snake_case : Union[str, Any] = atom_name if len(__lowercase ) == 4 else F""" {atom_name}""" _snake_case : str = "" _snake_case : str = "" _snake_case : Any = 1.00 _snake_case : str = atom_name[0] # Protein supports only C, N, O, S, this works. _snake_case : Dict = "" _snake_case : Any = "A" if chain_index is not None: _snake_case : List[Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! _snake_case : Optional[int] = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(__lowercase ) atom_index += 1 _snake_case : Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: _snake_case : Optional[int] = True _snake_case : Union[str, Any] = chain_index[i + 1] if should_terminate: # Close the chain. _snake_case : List[str] = "TER" _snake_case : str = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(__lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__lowercase , __lowercase ) ) pdb_lines.append("END" ) pdb_lines.append("" ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Protein: '''simple docstring''' return Protein( aatype=features["aatype"] , atom_positions=result["final_atom_positions"] , atom_mask=result["final_atom_mask"] , residue_index=features["residue_index"] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["final_atom_mask"] ) , chain_index=__lowercase , remark=__lowercase , parents=__lowercase , parents_chain_index=__lowercase , )
670
1
import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch if is_datasets_available(): from datasets import load_dataset __SCREAMING_SNAKE_CASE : Dict = random.Random() def snake_case (__lowercase , __lowercase=1.0 , __lowercase=None , __lowercase=None ) -> Union[str, Any]: '''simple docstring''' if rng is None: _snake_case : List[str] = global_rng _snake_case : Optional[int] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowercase_ ( unittest.TestCase ): def __init__( self , lowercase_ , lowercase_=7 , lowercase_=400 , lowercase_=2_000 , lowercase_=2_048 , lowercase_=128 , lowercase_=1 , lowercase_=512 , lowercase_=30 , lowercase_=44_100 , ): _snake_case : Optional[Any] = parent _snake_case : List[str] = batch_size _snake_case : List[str] = min_seq_length _snake_case : List[Any] = max_seq_length _snake_case : Dict = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _snake_case : Union[str, Any] = spectrogram_length _snake_case : Optional[int] = feature_size _snake_case : List[str] = num_audio_channels _snake_case : List[str] = hop_length _snake_case : int = chunk_length _snake_case : Tuple = sampling_rate def UpperCamelCase ( self ): return { "spectrogram_length": self.spectrogram_length, "feature_size": self.feature_size, "num_audio_channels": self.num_audio_channels, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "sampling_rate": self.sampling_rate, } def UpperCamelCase ( self , lowercase_=False , lowercase_=False ): def _flatten(lowercase_ ): return list(itertools.chain(*lowercase_ ) ) if equal_length: _snake_case : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _snake_case : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _snake_case : int = [np.asarray(lowercase_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = TvltFeatureExtractor def UpperCamelCase ( self ): _snake_case : Dict = TvltFeatureExtractionTester(self ) def UpperCamelCase ( self ): _snake_case : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(lowercase_ , "spectrogram_length" ) ) self.assertTrue(hasattr(lowercase_ , "feature_size" ) ) self.assertTrue(hasattr(lowercase_ , "num_audio_channels" ) ) self.assertTrue(hasattr(lowercase_ , "hop_length" ) ) self.assertTrue(hasattr(lowercase_ , "chunk_length" ) ) self.assertTrue(hasattr(lowercase_ , "sampling_rate" ) ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Dict = feat_extract_first.save_pretrained(lowercase_ )[0] check_json_file_has_correct_format(lowercase_ ) _snake_case : Optional[int] = self.feature_extraction_class.from_pretrained(lowercase_ ) _snake_case : Union[str, Any] = feat_extract_first.to_dict() _snake_case : Optional[int] = feat_extract_second.to_dict() _snake_case : List[str] = dict_first.pop("mel_filters" ) _snake_case : Optional[int] = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(lowercase_ , lowercase_ ) ) self.assertEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : int = os.path.join(lowercase_ , "feat_extract.json" ) feat_extract_first.to_json_file(lowercase_ ) _snake_case : Tuple = self.feature_extraction_class.from_json_file(lowercase_ ) _snake_case : Optional[int] = feat_extract_first.to_dict() _snake_case : Optional[int] = feat_extract_second.to_dict() _snake_case : Tuple = dict_first.pop("mel_filters" ) _snake_case : str = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(lowercase_ , lowercase_ ) ) self.assertEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): # Initialize feature_extractor _snake_case : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) # create three inputs of length 800, 1000, and 1200 _snake_case : str = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] _snake_case : int = [np.asarray(lowercase_ ) for speech_input in speech_inputs] # Test not batched input _snake_case : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test batched _snake_case : List[str] = feature_extractor(lowercase_ , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test audio masking _snake_case : Optional[int] = feature_extractor( lowercase_ , return_tensors="np" , sampling_rate=44_100 , mask_audio=lowercase_ ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test 2-D numpy arrays are batched. _snake_case : Union[str, Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] _snake_case : List[Any] = np.asarray(lowercase_ ) _snake_case : str = feature_extractor(lowercase_ , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) def UpperCamelCase ( self , lowercase_ ): _snake_case : Dict = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech _snake_case : Tuple = ds.sort("id" ).select(range(lowercase_ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def UpperCamelCase ( self ): _snake_case : Any = self._load_datasamples(1 ) _snake_case : List[Any] = TvltFeatureExtractor() _snake_case : Optional[Any] = feature_extractor(lowercase_ , return_tensors="pt" ).audio_values self.assertEquals(audio_values.shape , (1, 1, 192, 128) ) _snake_case : List[Any] = torch.tensor([[-0.3_032, -0.2_708], [-0.4_434, -0.4_007]] ) self.assertTrue(torch.allclose(audio_values[0, 0, :2, :2] , lowercase_ , atol=1e-4 ) )
670
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor'] _lowerCamelCase = 'SamImageProcessor' def __init__( self , lowercase_ ): super().__init__(lowercase_ ) _snake_case : Optional[Any] = self.image_processor _snake_case : Tuple = -10 _snake_case : str = self.image_processor.size["longest_edge"] def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_ = None , **lowercase_ , ): _snake_case : List[Any] = self.image_processor( lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # pop arguments that are not used in the foward but used nevertheless _snake_case : Any = encoding_image_processor["original_sizes"] if hasattr(lowercase_ , "numpy" ): # Checks if Torch or TF tensor _snake_case : int = original_sizes.numpy() _snake_case ,_snake_case ,_snake_case : Union[str, Any] = self._check_and_preprocess_points( input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , ) _snake_case : Dict = self._normalize_and_convert( lowercase_ , lowercase_ , input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , return_tensors=lowercase_ , ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_="pt" , ): if input_points is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : int = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] ) for point in input_points ] else: _snake_case : Dict = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ ) for point, original_size in zip(lowercase_ , lowercase_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _snake_case ,_snake_case : int = self._pad_points_and_labels(lowercase_ , lowercase_ ) _snake_case : Any = np.array(lowercase_ ) if input_labels is not None: _snake_case : Optional[Any] = np.array(lowercase_ ) if input_boxes is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : Optional[Any] = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] , is_bounding_box=lowercase_ ) for box in input_boxes ] else: _snake_case : List[str] = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ , is_bounding_box=lowercase_ ) for box, original_size in zip(lowercase_ , lowercase_ ) ] _snake_case : Tuple = np.array(lowercase_ ) if input_boxes is not None: if return_tensors == "pt": _snake_case : List[str] = torch.from_numpy(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[int] = tf.expand_dims(lowercase_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": _snake_case : Tuple = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : int = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : Tuple = tf.expand_dims(lowercase_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": _snake_case : Dict = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : str = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _snake_case : Optional[Any] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : List[Any] = tf.expand_dims(lowercase_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : List[Any] = max([point.shape[0] for point in input_points] ) _snake_case : List[str] = [] for i, point in enumerate(lowercase_ ): if point.shape[0] != expected_nb_points: _snake_case : Optional[Any] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _snake_case : Union[str, Any] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowercase_ ) _snake_case : Optional[Any] = processed_input_points return input_points, input_labels def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=False ): _snake_case ,_snake_case : Optional[int] = original_size _snake_case ,_snake_case : List[str] = self.image_processor._get_preprocess_shape(lowercase_ , longest_edge=lowercase_ ) _snake_case : Optional[Any] = deepcopy(lowercase_ ).astype(lowercase_ ) if is_bounding_box: _snake_case : str = coords.reshape(-1 , 2 , 2 ) _snake_case : Optional[Any] = coords[..., 0] * (new_w / old_w) _snake_case : Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: _snake_case : Optional[Any] = coords.reshape(-1 , 4 ) return coords def UpperCamelCase ( self , lowercase_=None , lowercase_=None , lowercase_=None , ): if input_points is not None: if hasattr(lowercase_ , "numpy" ): # Checks for TF or Torch tensor _snake_case : Union[str, Any] = input_points.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_points[0] , lowercase_ ): raise ValueError("Input points must be a list of list of floating points." ) _snake_case : Any = [np.array(lowercase_ ) for input_point in input_points] else: _snake_case : Optional[int] = None if input_labels is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : Tuple = input_labels.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_labels[0] , lowercase_ ): raise ValueError("Input labels must be a list of list integers." ) _snake_case : Tuple = [np.array(lowercase_ ) for label in input_labels] else: _snake_case : Optional[Any] = None if input_boxes is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : List[str] = input_boxes.numpy().tolist() if ( not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_boxes[0] , lowercase_ ) or not isinstance(input_boxes[0][0] , lowercase_ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) _snake_case : List[Any] = [np.array(lowercase_ ).astype(np.floataa ) for box in input_boxes] else: _snake_case : Optional[int] = None return input_points, input_labels, input_boxes @property def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(lowercase_ ) ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.image_processor.post_process_masks(*lowercase_ , **lowercase_ )
670
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[int] = { 'microsoft/trocr-base-handwritten': ( 'https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class lowercase_ ( __snake_case ): _lowerCamelCase = 'trocr' _lowerCamelCase = ['past_key_values'] _lowerCamelCase = { 'num_attention_heads': 'decoder_attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'decoder_layers', } def __init__( self , lowercase_=50_265 , lowercase_=1_024 , lowercase_=12 , lowercase_=16 , lowercase_=4_096 , lowercase_="gelu" , lowercase_=512 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=2 , lowercase_=0.02 , lowercase_=0.0 , lowercase_=True , lowercase_=False , lowercase_=True , lowercase_=True , lowercase_=1 , lowercase_=0 , lowercase_=2 , **lowercase_ , ): _snake_case : str = vocab_size _snake_case : List[str] = d_model _snake_case : Tuple = decoder_layers _snake_case : List[str] = decoder_attention_heads _snake_case : int = decoder_ffn_dim _snake_case : int = activation_function _snake_case : str = max_position_embeddings _snake_case : Any = dropout _snake_case : Any = attention_dropout _snake_case : Dict = activation_dropout _snake_case : str = init_std _snake_case : Union[str, Any] = decoder_layerdrop _snake_case : Any = use_cache _snake_case : str = scale_embedding _snake_case : Union[str, Any] = use_learned_position_embeddings _snake_case : Union[str, Any] = layernorm_embedding super().__init__( pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , decoder_start_token_id=lowercase_ , **lowercase_ , )
670
def snake_case (__lowercase ) -> int: '''simple docstring''' if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _snake_case : Union[str, Any] = grid[0] for row_n in range(1 , len(__lowercase ) ): _snake_case : Union[str, Any] = grid[row_n] _snake_case : List[Any] = fill_row(__lowercase , __lowercase ) _snake_case : List[Any] = grid[row_n] return grid[-1][-1] def snake_case (__lowercase , __lowercase ) -> list: '''simple docstring''' current_row[0] += row_above[0] for cell_n in range(1 , len(__lowercase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
670
1
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class lowercase_ ( unittest.TestCase ): _lowerCamelCase = MODEL_FOR_MASKED_LM_MAPPING _lowerCamelCase = TF_MODEL_FOR_MASKED_LM_MAPPING def UpperCamelCase ( self ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def UpperCamelCase ( self ): _snake_case : List[Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf" ) _snake_case : Dict = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ {"sequence": "My name is grouped", "score": 2.1e-05, "token": 38_015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-05, "token": 25_506, "token_str": " accuser"}, ] , ) _snake_case : Any = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-05, "token": 38_015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-05, "token": 25_506, "token_str": " accuser", }, ] , ) _snake_case : int = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ {"sequence": "My name is Clara", "score": 2e-05, "token": 13_606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-05, "token": 3_499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-05, "token": 2_941, "token_str": " Te"}, ] , ) @require_torch def UpperCamelCase ( self ): _snake_case : Optional[Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt" ) _snake_case : int = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ {"sequence": "My name is Maul", "score": 2.2e-05, "token": 35_676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-05, "token": 16_416, "token_str": "ELS"}, ] , ) _snake_case : str = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-05, "token": 35_676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-05, "token": 16_416, "token_str": "ELS"}, ] , ) _snake_case : Optional[Any] = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ {"sequence": "My name is Patrick", "score": 2.1e-05, "token": 3_499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-05, "token": 2_941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-05, "token": 13_606, "token_str": " Clara"}, ] , ) _snake_case : Dict = unmasker("My name is <mask> <mask>" , top_k=2 ) self.assertEqual( nested_simplify(lowercase_ , decimals=6 ) , [ [ { "score": 2.2e-05, "token": 35_676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-05, "token": 16_416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-05, "token": 35_676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-05, "token": 16_416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def UpperCamelCase ( self ): _snake_case : Optional[int] = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt" ) # convert model to fp16 pipe.model.half() _snake_case : List[Any] = pipe("Paris is the [MASK] of France." ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase_ , lowercase_ ) @slow @require_torch def UpperCamelCase ( self ): _snake_case : int = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt" ) self.run_large_test(lowercase_ ) @slow @require_tf def UpperCamelCase ( self ): _snake_case : Union[str, Any] = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf" ) self.run_large_test(lowercase_ ) def UpperCamelCase ( self , lowercase_ ): _snake_case : Optional[int] = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowercase_ ) , [ {"sequence": "My name is John", "score": 0.008, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.007, "token": 1_573, "token_str": " Chris"}, ] , ) _snake_case : Optional[Any] = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowercase_ ) , [ { "sequence": "The largest city in France is Paris", "score": 0.251, "token": 2_201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.214, "token": 12_790, "token_str": " Lyon", }, ] , ) _snake_case : str = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowercase_ ) , [ {"sequence": "My name is Patrick", "score": 0.005, "token": 3_499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.000, "token": 13_606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.000, "token": 2_941, "token_str": " Te"}, ] , ) @require_torch def UpperCamelCase ( self ): _snake_case : str = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt" ) _snake_case : Tuple = None _snake_case : Optional[Any] = None self.run_pipeline_test(lowercase_ , [] ) @require_tf def UpperCamelCase ( self ): _snake_case : List[str] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf" ) _snake_case : int = None _snake_case : str = None self.run_pipeline_test(lowercase_ , [] ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ ): if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)" ) _snake_case : Union[str, Any] = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) _snake_case : str = [ f"""This is another {tokenizer.mask_token} test""", ] return fill_masker, examples def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Tuple = fill_masker.tokenizer _snake_case : int = fill_masker.model _snake_case : Tuple = fill_masker( f"""This is a {tokenizer.mask_token}""" , ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) _snake_case : Any = fill_masker([f"""This is a {tokenizer.mask_token}"""] ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) _snake_case : str = fill_masker([f"""This is a {tokenizer.mask_token}""", f"""Another {tokenizer.mask_token} great test."""] ) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ], [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ], ] , ) with self.assertRaises(lowercase_ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(lowercase_ ): fill_masker("This is" ) self.run_test_top_k(lowercase_ , lowercase_ ) self.run_test_targets(lowercase_ , lowercase_ ) self.run_test_top_k_targets(lowercase_ , lowercase_ ) self.fill_mask_with_duplicate_targets_and_top_k(lowercase_ , lowercase_ ) self.fill_mask_with_multiple_masks(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Optional[int] = tokenizer.get_vocab() _snake_case : Optional[int] = sorted(vocab.keys() )[:2] # Pipeline argument _snake_case : Dict = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , targets=lowercase_ ) _snake_case : Dict = fill_masker(f"""This is a {tokenizer.mask_token}""" ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) _snake_case : Union[str, Any] = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_ ) _snake_case : Any = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_ ) ) # Call argument _snake_case : Optional[Any] = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) _snake_case : Tuple = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=lowercase_ ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) _snake_case : str = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowercase_ ) _snake_case : Optional[Any] = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowercase_ ) ) # Score equivalence _snake_case : str = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=lowercase_ ) _snake_case : Optional[int] = [top_mask["token_str"] for top_mask in outputs] _snake_case : List[Any] = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_ ) == set(lowercase_ ): _snake_case : str = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=lowercase_ ) _snake_case : Tuple = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase_ ) , nested_simplify(lowercase_ ) ) # Raises with invalid with self.assertRaises(lowercase_ ): _snake_case : Optional[int] = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase_ ): _snake_case : Union[str, Any] = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=[""] ) with self.assertRaises(lowercase_ ): _snake_case : Optional[int] = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets="" ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Optional[int] = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ , top_k=2 ) _snake_case : Dict = fill_masker(f"""This is a {tokenizer.mask_token}""" ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) _snake_case : Tuple = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) _snake_case : Optional[int] = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( lowercase_ , [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ] , ) self.assertEqual(nested_simplify(lowercase_ ) , nested_simplify(lowercase_ ) ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Tuple = tokenizer.get_vocab() _snake_case : Any = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) # top_k=2, ntargets=3 _snake_case : Union[str, Any] = sorted(vocab.keys() )[:3] _snake_case : Tuple = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=2 , targets=lowercase_ ) # If we use the most probably targets, and filter differently, we should still # have the same results _snake_case : str = [el["token_str"] for el in sorted(lowercase_ , key=lambda lowercase_ : x["score"] , reverse=lowercase_ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase_ ).issubset(lowercase_ ): _snake_case : Any = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=3 , targets=lowercase_ ) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase_ ) , nested_simplify(lowercase_ ) ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : int = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) _snake_case : List[Any] = tokenizer.get_vocab() # String duplicates + id duplicates _snake_case : Tuple = sorted(vocab.keys() )[:3] _snake_case : Optional[int] = [targets[0], targets[1], targets[0], targets[2], targets[1]] _snake_case : Dict = fill_masker(f"""My name is {tokenizer.mask_token}""" , targets=lowercase_ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase_ ) , 3 ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Union[str, Any] = FillMaskPipeline(model=lowercase_ , tokenizer=lowercase_ ) _snake_case : Optional[int] = fill_masker( f"""This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( lowercase_ , [ [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ], [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ], [ {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, {"sequence": ANY(lowercase_ ), "score": ANY(lowercase_ ), "token": ANY(lowercase_ ), "token_str": ANY(lowercase_ )}, ], ] , )
670
import random def snake_case (__lowercase , __lowercase ) -> tuple: '''simple docstring''' _snake_case ,_snake_case ,_snake_case : List[Any] = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def snake_case (__lowercase , __lowercase ) -> List[Any]: '''simple docstring''' if index >= len(__lowercase ) or index < 0: return None _snake_case : Any = items[random.randint(0 , len(__lowercase ) - 1 )] _snake_case : Tuple = 0 _snake_case ,_snake_case ,_snake_case : Tuple = _partition(__lowercase , __lowercase ) _snake_case : Tuple = len(__lowercase ) _snake_case : List[str] = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
670
1
import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DonutImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
from math import pow, sqrt def snake_case (*__lowercase ) -> bool: '''simple docstring''' _snake_case : str = len(__lowercase ) > 0 and all(value > 0.0 for value in values ) return result def snake_case (__lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
670
1
def snake_case (__lowercase , __lowercase , __lowercase , __lowercase ) -> Tuple: '''simple docstring''' if height >= 1: move_tower(height - 1 , __lowercase , __lowercase , __lowercase ) move_disk(__lowercase , __lowercase ) move_tower(height - 1 , __lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase ) -> Tuple: '''simple docstring''' print("moving disk from" , __lowercase , "to" , __lowercase ) def snake_case () -> Union[str, Any]: '''simple docstring''' _snake_case : int = int(input("Height of hanoi: " ).strip() ) move_tower(__lowercase , "A" , "B" , "C" ) if __name__ == "__main__": main()
670
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
1
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
from __future__ import annotations from typing import TypedDict class lowercase_ ( __snake_case ): _lowerCamelCase = 42 _lowerCamelCase = 42 def snake_case (__lowercase ) -> list[str]: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(__lowercase ) )] def snake_case (__lowercase ) -> BWTTransformDict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) _snake_case : List[str] = all_rotations(__lowercase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation _snake_case : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__lowercase ), } return response def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: _snake_case : Union[str, Any] = int(__lowercase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(__lowercase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) _snake_case : Optional[Any] = [""] * len(__lowercase ) for _ in range(len(__lowercase ) ): for i in range(len(__lowercase ) ): _snake_case : Tuple = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = 'Provide a string that I will generate its BWT transform: ' __SCREAMING_SNAKE_CASE : Optional[Any] = input(entry_msg).strip() __SCREAMING_SNAKE_CASE : int = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result['bwt_string']}\'''' ) __SCREAMING_SNAKE_CASE : List[str] = reverse_bwt(result['bwt_string'], result['idx_original_string']) print( F'''Reversing Burrows Wheeler transform for entry \'{result['bwt_string']}\' ''' F'''we get original string \'{original_string}\'''' )
670
1
def snake_case (__lowercase , __lowercase ) -> Union[str, Any]: '''simple docstring''' return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def snake_case (__lowercase , __lowercase=0 ) -> Optional[Any]: '''simple docstring''' return sorted(__lowercase , key=lambda __lowercase : x[column] ) def snake_case (__lowercase , __lowercase , __lowercase=float("inf" ) ) -> str: '''simple docstring''' for i in range(points_counts - 1 ): for j in range(i + 1 , __lowercase ): _snake_case : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: _snake_case : Optional[int] = current_dis return min_dis def snake_case (__lowercase , __lowercase , __lowercase=float("inf" ) ) -> Dict: '''simple docstring''' for i in range(min(6 , points_counts - 1 ) , __lowercase ): for j in range(max(0 , i - 6 ) , __lowercase ): _snake_case : int = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: _snake_case : Any = current_dis return min_dis def snake_case (__lowercase , __lowercase , __lowercase ) -> Any: '''simple docstring''' if points_counts <= 3: return dis_between_closest_pair(__lowercase , __lowercase ) # recursion _snake_case : List[str] = points_counts // 2 _snake_case : Optional[Any] = closest_pair_of_points_sqr( __lowercase , points_sorted_on_y[:mid] , __lowercase ) _snake_case : str = closest_pair_of_points_sqr( __lowercase , points_sorted_on_y[mid:] , points_counts - mid ) _snake_case : Union[str, Any] = min(__lowercase , __lowercase ) _snake_case : Any = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(__lowercase ) _snake_case : List[str] = dis_between_closest_in_strip( __lowercase , len(__lowercase ) , __lowercase ) return min(__lowercase , __lowercase ) def snake_case (__lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[int] = column_based_sort(__lowercase , column=0 ) _snake_case : int = column_based_sort(__lowercase , column=1 ) return ( closest_pair_of_points_sqr( __lowercase , __lowercase , __lowercase ) ) ** 0.5 if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = [(2, 3), (1_2, 3_0), (4_0, 5_0), (5, 1), (1_2, 1_0), (3, 4)] print('Distance:', closest_pair_of_points(points, len(points)))
670
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( 'stable diffusion controlnet', '0.22.0', 'Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.', standard_warn=False, stacklevel=3, )
670
1
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __SCREAMING_SNAKE_CASE : List[str] = Mapping[str, np.ndarray] __SCREAMING_SNAKE_CASE : List[Any] = Mapping[str, Any] # Is a nested dict. __SCREAMING_SNAKE_CASE : List[Any] = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class lowercase_ : _lowerCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _lowerCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _lowerCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _lowerCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _lowerCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _lowerCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _lowerCamelCase = None # Templates used to generate this protein (prediction-only) _lowerCamelCase = None # Chain corresponding to each parent _lowerCamelCase = None def snake_case (__lowercase ) -> Protein: '''simple docstring''' _snake_case : str = r"(\[[A-Z]+\]\n)" _snake_case : List[str] = [tag.strip() for tag in re.split(__lowercase , __lowercase ) if len(__lowercase ) > 0] _snake_case : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("\n" ) for l in tags[1::2]] ) _snake_case : List[str] = ["N", "CA", "C"] _snake_case : Any = None _snake_case : Union[str, Any] = None _snake_case : Optional[int] = None for g in groups: if "[PRIMARY]" == g[0]: _snake_case : Tuple = g[1][0].strip() for i in range(len(__lowercase ) ): if seq[i] not in residue_constants.restypes: _snake_case : Tuple = "X" # FIXME: strings are immutable _snake_case : int = np.array( [residue_constants.restype_order.get(__lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: _snake_case : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(__lowercase , g[1][axis].split() ) ) ) _snake_case : Dict = np.array(__lowercase ) _snake_case : Dict = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: _snake_case : int = np.array(list(map({"-": 0, "+": 1}.get , g[1][0].strip() ) ) ) _snake_case : Any = np.zeros( ( len(__lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : Dict = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__lowercase , atom_mask=__lowercase , aatype=__lowercase , residue_index=np.arange(len(__lowercase ) ) , b_factors=__lowercase , ) def snake_case (__lowercase , __lowercase = 0 ) -> List[str]: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[Any] = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) _snake_case : str = prot.parents _snake_case : str = prot.parents_chain_index if parents is not None and parents_chain_index is not None: _snake_case : int = [p for i, p in zip(__lowercase , __lowercase ) if i == chain_id] if parents is None or len(__lowercase ) == 0: _snake_case : Optional[int] = ["N/A"] pdb_headers.append(F"""PARENT {' '.join(__lowercase )}""" ) return pdb_headers def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[int] = pdb_str.split("\n" ) _snake_case : List[str] = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) _snake_case : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: _snake_case : str = [] if prot.parents_chain_index is not None: _snake_case : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(__lowercase ) , [] ) parent_dict[str(__lowercase )].append(__lowercase ) _snake_case : Any = max([int(__lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): _snake_case : Tuple = parent_dict.get(str(__lowercase ) , ["N/A"] ) parents_per_chain.append(__lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: _snake_case : List[str] = [["N/A"]] def make_parent_line(__lowercase ) -> str: return F"""PARENT {' '.join(__lowercase )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) _snake_case : int = 0 for i, l in enumerate(__lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__lowercase ): _snake_case : Tuple = parents_per_chain[chain_counter] else: _snake_case : str = ["N/A"] out_pdb_lines.append(make_parent_line(__lowercase ) ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = residue_constants.restypes + ["X"] def res_atoa(__lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , "UNK" ) _snake_case : Optional[int] = residue_constants.atom_types _snake_case : List[str] = [] _snake_case : Tuple = prot.atom_mask _snake_case : List[str] = prot.aatype _snake_case : int = prot.atom_positions _snake_case : int = prot.residue_index.astype(np.intaa ) _snake_case : List[Any] = prot.b_factors _snake_case : str = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("Invalid aatypes." ) _snake_case : Union[str, Any] = get_pdb_headers(__lowercase ) if len(__lowercase ) > 0: pdb_lines.extend(__lowercase ) _snake_case : Optional[Any] = aatype.shape[0] _snake_case : str = 1 _snake_case : Tuple = 0 _snake_case : int = string.ascii_uppercase _snake_case : Optional[Any] = None # Add all atom sites. for i in range(__lowercase ): _snake_case : Dict = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue _snake_case : List[Any] = "ATOM" _snake_case : Union[str, Any] = atom_name if len(__lowercase ) == 4 else F""" {atom_name}""" _snake_case : str = "" _snake_case : str = "" _snake_case : Any = 1.00 _snake_case : str = atom_name[0] # Protein supports only C, N, O, S, this works. _snake_case : Dict = "" _snake_case : Any = "A" if chain_index is not None: _snake_case : List[Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! _snake_case : Optional[int] = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(__lowercase ) atom_index += 1 _snake_case : Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: _snake_case : Optional[int] = True _snake_case : Union[str, Any] = chain_index[i + 1] if should_terminate: # Close the chain. _snake_case : List[str] = "TER" _snake_case : str = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(__lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__lowercase , __lowercase ) ) pdb_lines.append("END" ) pdb_lines.append("" ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Protein: '''simple docstring''' return Protein( aatype=features["aatype"] , atom_positions=result["final_atom_positions"] , atom_mask=result["final_atom_mask"] , residue_index=features["residue_index"] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["final_atom_mask"] ) , chain_index=__lowercase , remark=__lowercase , parents=__lowercase , parents_chain_index=__lowercase , )
670
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase_ : _lowerCamelCase = LEDConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=4 , ): _snake_case : Optional[int] = parent _snake_case : str = batch_size _snake_case : int = seq_length _snake_case : Dict = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = vocab_size _snake_case : str = hidden_size _snake_case : int = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : int = intermediate_size _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : int = max_position_embeddings _snake_case : Union[str, Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Any = bos_token_id _snake_case : str = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after _snake_case : List[Any] = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests _snake_case : List[str] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def UpperCamelCase ( self ): _snake_case : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Any = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) _snake_case : Optional[Any] = prepare_led_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) _snake_case : int = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]] , axis=-1 , ) _snake_case : List[Any] = global_attention_mask return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = TFLEDModel(config=lowercase_ ).get_decoder() _snake_case : Optional[Any] = inputs_dict["input_ids"] _snake_case : Optional[int] = input_ids[:1, :] _snake_case : int = inputs_dict["attention_mask"][:1, :] _snake_case : int = 1 # first forward pass _snake_case : str = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case : Tuple = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case : List[str] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case : str = model(lowercase_ , attention_mask=lowercase_ )[0] _snake_case : List[str] = model(lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_ , lowercase_ , rtol=1e-3 ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> List[Any]: '''simple docstring''' if attention_mask is None: _snake_case : int = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () _lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFLEDModelTester(self ) _snake_case : List[Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = tf.zeros_like(inputs_dict["attention_mask"] ) _snake_case : Tuple = 2 _snake_case : Dict = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) _snake_case : Tuple = True _snake_case : Union[str, Any] = self.model_tester.seq_length _snake_case : Union[str, Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ ): _snake_case : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(lowercase_ ): _snake_case : int = [t.numpy() for t in outputs.encoder_attentions] _snake_case : Optional[int] = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: _snake_case : Union[str, Any] = True _snake_case : Dict = False _snake_case : Any = False _snake_case : Any = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) _snake_case : Tuple = len(lowercase_ ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: _snake_case : int = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Any = model_class(lowercase_ ) _snake_case : Optional[Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine _snake_case : Optional[int] = True _snake_case : Optional[int] = True _snake_case : List[Any] = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # TODO: Head-masking not yet implement pass def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' return tf.constant(__lowercase , dtype=tf.intaa ) __SCREAMING_SNAKE_CASE : List[Any] = 1E-4 @slow @require_tf class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Dict = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here _snake_case : Union[str, Any] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[int] = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Union[str, Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Optional[Any] = model(**lowercase_ )[0] _snake_case : str = (1, 1_024, 768) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[Any] = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 ) def UpperCamelCase ( self ): _snake_case : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here _snake_case : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : int = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Tuple = model(**lowercase_ )[0] _snake_case : Any = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[int] = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 , rtol=1e-3 )
670
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __SCREAMING_SNAKE_CASE : Any = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = ['GPTSw3Tokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
1
def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : list[list[str]] = [[] for _ in range(__lowercase )] _snake_case : int = key - 1 if key <= 0: raise ValueError("Height of grid can't be 0 or negative" ) if key == 1 or len(__lowercase ) <= key: return input_string for position, character in enumerate(__lowercase ): _snake_case : str = position % (lowest * 2) # puts it in bounds _snake_case : Any = min(__lowercase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(__lowercase ) _snake_case : Optional[int] = ["".join(__lowercase ) for row in temp_grid] _snake_case : Tuple = "".join(__lowercase ) return output_string def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : Dict = [] _snake_case : Union[str, Any] = key - 1 if key <= 0: raise ValueError("Height of grid can't be 0 or negative" ) if key == 1: return input_string _snake_case : list[list[str]] = [[] for _ in range(__lowercase )] # generates template for position in range(len(__lowercase ) ): _snake_case : Optional[Any] = position % (lowest * 2) # puts it in bounds _snake_case : int = min(__lowercase , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("*" ) _snake_case : List[str] = 0 for row in temp_grid: # fills in the characters _snake_case : Union[str, Any] = input_string[counter : counter + len(__lowercase )] grid.append(list(__lowercase ) ) counter += len(__lowercase ) _snake_case : Optional[Any] = "" # reads as zigzag for position in range(len(__lowercase ) ): _snake_case : Tuple = position % (lowest * 2) # puts it in bounds _snake_case : int = min(__lowercase , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case (__lowercase ) -> dict[int, str]: '''simple docstring''' _snake_case : Optional[Any] = {} for key_guess in range(1 , len(__lowercase ) ): # tries every key _snake_case : Tuple = decrypt(__lowercase , __lowercase ) return results if __name__ == "__main__": import doctest doctest.testmod()
670
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Any = tempfile.mkdtemp() # fmt: off _snake_case : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _snake_case : Dict = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] _snake_case : Optional[int] = {"unk_token": "<unk>"} _snake_case : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) _snake_case : Any = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } _snake_case : Optional[Any] = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Union[str, Any] = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Tuple = self.get_tokenizer() _snake_case : Any = self.get_rust_tokenizer() _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) _snake_case : Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) _snake_case : List[Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : List[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Tuple = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : int = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Optional[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : str = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : Dict = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : int = processor(text=lowercase_ ) _snake_case : str = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[Any] = "lower newer" _snake_case : int = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[str] = self.get_tokenizer() _snake_case : Union[str, Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Dict = self.prepare_image_inputs() _snake_case : List[Any] = processor(images=lowercase_ , visual_prompt=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[Any] = self.get_tokenizer() _snake_case : str = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Any = processor.batch_decode(lowercase_ ) _snake_case : Any = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ )
670
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { 'SenseTime/deformable-detr': 'https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowercase_ ( __snake_case ): _lowerCamelCase = 'deformable_detr' _lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=300 , lowercase_=1_024 , lowercase_=6 , lowercase_=1_024 , lowercase_=8 , lowercase_=6 , lowercase_=1_024 , lowercase_=8 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=True , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=4 , lowercase_=4 , lowercase_=4 , lowercase_=False , lowercase_=300 , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , lowercase_=0.25 , lowercase_=False , **lowercase_ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) _snake_case : Dict = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): _snake_case : str = backbone_config.get("model_type" ) _snake_case : List[str] = CONFIG_MAPPING[backbone_model_type] _snake_case : Optional[int] = config_class.from_dict(lowercase_ ) _snake_case : List[str] = use_timm_backbone _snake_case : Union[str, Any] = backbone_config _snake_case : List[str] = num_channels _snake_case : List[str] = num_queries _snake_case : int = max_position_embeddings _snake_case : Optional[Any] = d_model _snake_case : int = encoder_ffn_dim _snake_case : str = encoder_layers _snake_case : str = encoder_attention_heads _snake_case : str = decoder_ffn_dim _snake_case : Union[str, Any] = decoder_layers _snake_case : List[str] = decoder_attention_heads _snake_case : int = dropout _snake_case : Optional[int] = attention_dropout _snake_case : Dict = activation_dropout _snake_case : Union[str, Any] = activation_function _snake_case : List[str] = init_std _snake_case : Tuple = init_xavier_std _snake_case : str = encoder_layerdrop _snake_case : Optional[int] = auxiliary_loss _snake_case : List[str] = position_embedding_type _snake_case : Optional[int] = backbone _snake_case : Optional[Any] = use_pretrained_backbone _snake_case : Optional[Any] = dilation # deformable attributes _snake_case : List[Any] = num_feature_levels _snake_case : Dict = encoder_n_points _snake_case : List[str] = decoder_n_points _snake_case : Optional[int] = two_stage _snake_case : Dict = two_stage_num_proposals _snake_case : Tuple = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher _snake_case : List[str] = class_cost _snake_case : Optional[int] = bbox_cost _snake_case : Dict = giou_cost # Loss coefficients _snake_case : Dict = mask_loss_coefficient _snake_case : Tuple = dice_loss_coefficient _snake_case : List[str] = bbox_loss_coefficient _snake_case : Tuple = giou_loss_coefficient _snake_case : List[Any] = eos_coefficient _snake_case : List[str] = focal_alpha _snake_case : Optional[Any] = disable_custom_kernels super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): return self.encoder_attention_heads @property def UpperCamelCase ( self ): return self.d_model def UpperCamelCase ( self ): _snake_case : str = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _snake_case : Tuple = self.backbone_config.to_dict() _snake_case : Tuple = self.__class__.model_type return output
670
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
1
__SCREAMING_SNAKE_CASE : List[str] = 'Alexander Joslin' import operator as op from .stack import Stack def snake_case (__lowercase ) -> int: '''simple docstring''' _snake_case : Optional[Any] = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} _snake_case : Stack[int] = Stack() _snake_case : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowercase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowercase ) elif i == ")": # RULE 4 _snake_case : Union[str, Any] = operator_stack.peek() operator_stack.pop() _snake_case : Tuple = operand_stack.peek() operand_stack.pop() _snake_case : List[Any] = operand_stack.peek() operand_stack.pop() _snake_case : Optional[int] = operators[opr](__lowercase , __lowercase ) operand_stack.push(__lowercase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = '(5 + ((4 * 2) * (2 + 3)))' # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
670
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ ): super().__init__() _snake_case : List[str] = nn.ModuleList(lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = True , ): for i, (image, scale, controlnet) in enumerate(zip(lowercase_ , lowercase_ , self.nets ) ): _snake_case ,_snake_case : Optional[int] = controlnet( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) # merge samples if i == 0: _snake_case ,_snake_case : Tuple = down_samples, mid_sample else: _snake_case : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase_ , lowercase_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , ): _snake_case : Tuple = 0 _snake_case : Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase_ , is_main_process=lowercase_ , save_function=lowercase_ , safe_serialization=lowercase_ , variant=lowercase_ , ) idx += 1 _snake_case : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : List[str] = 0 _snake_case : Optional[Any] = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case : Optional[Any] = pretrained_model_path while os.path.isdir(lowercase_ ): _snake_case : int = ControlNetModel.from_pretrained(lowercase_ , **lowercase_ ) controlnets.append(lowercase_ ) idx += 1 _snake_case : str = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(lowercase_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(lowercase_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(lowercase_ )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(lowercase_ )
670
1
def snake_case (__lowercase ) -> list: '''simple docstring''' _snake_case : Optional[Any] = len(__lowercase ) for i in range(1 , __lowercase ): _snake_case : Tuple = collection[i] _snake_case : Dict = 0 _snake_case : Optional[Any] = i - 1 while low <= high: _snake_case : Tuple = (low + high) // 2 if val < collection[mid]: _snake_case : Optional[int] = mid - 1 else: _snake_case : int = mid + 1 for j in range(__lowercase , __lowercase , -1 ): _snake_case : List[Any] = collection[j - 1] _snake_case : List[Any] = val return collection if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = input('Enter numbers separated by a comma:\n').strip() __SCREAMING_SNAKE_CASE : int = [int(item) for item in user_input.split(',')] print(binary_insertion_sort(unsorted))
670
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'CLIPImageProcessor' _lowerCamelCase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Dict = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _snake_case : str = self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if images is not None: _snake_case : List[str] = self.image_processor(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if text is not None and images is not None: _snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase_ ) , tensor_type=lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
1
def snake_case (__lowercase ) -> list[int]: '''simple docstring''' _snake_case : Union[str, Any] = [0 for i in range(len(__lowercase ) )] # initialize interval's left pointer and right pointer _snake_case ,_snake_case : List[str] = 0, 0 for i in range(1 , len(__lowercase ) ): # case when current index is inside the interval if i <= right_pointer: _snake_case : str = min(right_pointer - i + 1 , z_result[i - left_pointer] ) _snake_case : str = min_edge while go_next(__lowercase , __lowercase , __lowercase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: _snake_case ,_snake_case : Dict = i, i + z_result[i] - 1 return z_result def snake_case (__lowercase , __lowercase , __lowercase ) -> bool: '''simple docstring''' return i + z_result[i] < len(__lowercase ) and s[z_result[i]] == s[i + z_result[i]] def snake_case (__lowercase , __lowercase ) -> int: '''simple docstring''' _snake_case : Dict = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string _snake_case : Dict = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(__lowercase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
670
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
670
1
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __SCREAMING_SNAKE_CASE : int = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F'''{bindir}/../../examples/pytorch/translation'''): from run_translation import main # noqa set_seed(4_2) __SCREAMING_SNAKE_CASE : Optional[int] = 'sshleifer/student_marian_en_ro_6_1' __SCREAMING_SNAKE_CASE : int = 'sshleifer/tiny-mbart' @require_torch class lowercase_ ( __snake_case ): def UpperCamelCase ( self , lowercase_=False , lowercase_=None , lowercase_=True , lowercase_=True , lowercase_=True , lowercase_=True , ): _snake_case : Union[str, Any] = self.run_trainer( eval_steps=1 , max_len=12 , model_name=lowercase_ , num_train_epochs=1 , distributed=lowercase_ , extra_args_str=lowercase_ , predict_with_generate=lowercase_ , do_train=lowercase_ , do_eval=lowercase_ , do_predict=lowercase_ , ) _snake_case : str = TrainerState.load_from_json(os.path.join(lowercase_ , "trainer_state.json" ) ).log_history if not do_eval: return _snake_case : str = [log for log in logs if "eval_loss" in log.keys()] _snake_case : int = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats _snake_case : int = eval_metrics[-1] assert isinstance(last_step_stats["eval_bleu"] , lowercase_ ) assert not math.isnan(float(last_step_stats["eval_loss"] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def UpperCamelCase ( self ): self.run_seqaseq_quick() @require_torch_multi_gpu def UpperCamelCase ( self ): self.run_seqaseq_quick(distributed=lowercase_ ) @require_torch_multi_gpu def UpperCamelCase ( self ): self.run_seqaseq_quick(distributed=lowercase_ ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase ( self ): self.run_seqaseq_quick(distributed=lowercase_ , extra_args_str="--sharded_ddp simple" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase ( self ): self.run_seqaseq_quick(distributed=lowercase_ , extra_args_str="--sharded_ddp simple --fp16" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase ( self ): self.run_seqaseq_quick(distributed=lowercase_ , extra_args_str="--sharded_ddp zero_dp_2" , predict_with_generate=lowercase_ ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase ( self ): self.run_seqaseq_quick( distributed=lowercase_ , extra_args_str="--sharded_ddp zero_dp_2 --fp16" , predict_with_generate=lowercase_ ) @require_apex @require_torch_gpu def UpperCamelCase ( self ): # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=lowercase_ , extra_args_str="--fp16 --fp16_backend=apex" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowercase_ , extra_args_str="--fp16 --fp16_backend=apex" ) @parameterized.expand(["base", "low", "high", "mixed"] ) @require_torch_multi_gpu def UpperCamelCase ( self , lowercase_ ): # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout _snake_case : Tuple = { # test with the default log_level - should be info and thus log info once "base": {"extra_args_str": "", "n_matches": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes "low": {"extra_args_str": "--log_level debug --log_level_replica debug", "n_matches": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica "high": {"extra_args_str": "--log_level error --log_level_replica debug", "n_matches": 1}, # test with high log_level and log_level_replica - should be quiet on all processes "mixed": {"extra_args_str": "--log_level error --log_level_replica error", "n_matches": 0}, } _snake_case : List[str] = experiments[experiment_id] _snake_case : Dict = {"distributed": True, "predict_with_generate": False, "do_eval": False, "do_predict": False} _snake_case : Any = "Running training" with CaptureStderr() as cl: self.run_seqaseq_quick(**lowercase_ , extra_args_str=data["extra_args_str"] ) _snake_case : Any = len(re.findall(lowercase_ , cl.err ) ) self.assertEqual(lowercase_ , data["n_matches"] ) @slow def UpperCamelCase ( self ): _snake_case : str = self.run_trainer( eval_steps=2 , max_len=128 , model_name=lowercase_ , learning_rate=3e-4 , num_train_epochs=10 , distributed=lowercase_ , ) # Check metrics _snake_case : Tuple = TrainerState.load_from_json(os.path.join(lowercase_ , "trainer_state.json" ) ).log_history _snake_case : List[Any] = [log for log in logs if "eval_loss" in log.keys()] _snake_case : int = eval_metrics[0] _snake_case : str = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["eval_bleu"] , lowercase_ ) # test if do_predict saves generations and metrics _snake_case : str = os.listdir(lowercase_ ) _snake_case : List[Any] = {os.path.basename(lowercase_ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def UpperCamelCase ( self ): from transformers.training_args import OptimizerNames def train_and_return_metrics(lowercase_ ) -> Tuple[int, float]: _snake_case : List[str] = "--skip_memory_metrics 0" _snake_case : List[Any] = self.run_trainer( max_len=128 , model_name=lowercase_ , learning_rate=3e-4 , num_train_epochs=1 , optim=lowercase_ , distributed=lowercase_ , extra_args_str=lowercase_ , do_eval=lowercase_ , do_predict=lowercase_ , n_gpus_to_use=1 , ) # Check metrics _snake_case : Optional[Any] = TrainerState.load_from_json(Path(lowercase_ , "trainer_state.json" ) ).log_history _snake_case : Union[str, Any] = int(logs[0]["train_mem_gpu_peaked_delta"] / 2**20 ) _snake_case : List[str] = int(logs[0]["train_mem_gpu_alloc_delta"] / 2**20 ) _snake_case : Any = logs[0]["train_loss"] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss _snake_case ,_snake_case ,_snake_case : Union[str, Any] = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) _snake_case ,_snake_case ,_snake_case : Dict = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) _snake_case : Optional[Any] = gpu_alloc_mem_orig - gpu_alloc_mem_bnb _snake_case : Optional[Any] = gpu_peak_mem_orig + gpu_alloc_mem_orig _snake_case : Dict = gpu_peak_mem_bnb + gpu_alloc_mem_bnb _snake_case : int = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings _snake_case : List[str] = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowercase_ , lowercase_ , "should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got" f""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" f""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( lowercase_ , lowercase_ , "should use ~150MB less total gpu memory with BNB, compared to without it for this model but got" f""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" f""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( lowercase_ , lowercase_ , f"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = 3e-3 , lowercase_ = "adafactor" , lowercase_ = False , lowercase_ = None , lowercase_ = 0 , lowercase_ = True , lowercase_ = True , lowercase_ = True , lowercase_ = True , lowercase_ = None , ): _snake_case : Dict = self.test_file_dir / "../fixtures/tests_samples/wmt_en_ro" _snake_case : Optional[Any] = self.get_auto_remove_tmp_dir() _snake_case : int = f""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(lowercase_ )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(lowercase_ )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() _snake_case : List[str] = f""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(lowercase_ )} """.split() _snake_case : List[str] = "\n --do_predict\n ".split() _snake_case : Tuple = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: _snake_case : List[str] = get_gpu_count() _snake_case : Tuple = get_torch_dist_unique_port() _snake_case : List[str] = f""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() _snake_case : Any = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowercase_ , env=self.get_env() ) else: _snake_case : List[str] = ["run_translation.py"] + args with patch.object(lowercase_ , "argv" , lowercase_ ): main() return output_dir
670
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def snake_case (*__lowercase ) -> Dict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): _snake_case : Dict = list(__lowercase ) for i in range(len(__lowercase ) ): _snake_case : List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : str = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowercase , __lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def snake_case (__lowercase = None , __lowercase = 128 ) -> Any: '''simple docstring''' if function is None: return functools.partial(__lowercase , starting_batch_size=__lowercase ) _snake_case : List[str] = starting_batch_size def decorator(*__lowercase , **__lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _snake_case : Optional[Any] = list(inspect.signature(__lowercase ).parameters.keys() ) # Guard against user error if len(__lowercase ) < (len(__lowercase ) + 1): _snake_case : str = ", ".join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowercase , *__lowercase , **__lowercase ) except Exception as e: if should_reduce_batch_size(__lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
670
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'vocab_file': 'vocab.txt'} __SCREAMING_SNAKE_CASE : List[Any] = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } __SCREAMING_SNAKE_CASE : str = { 'YituTech/conv-bert-base': 5_1_2, 'YituTech/conv-bert-medium-small': 5_1_2, 'YituTech/conv-bert-small': 5_1_2, } __SCREAMING_SNAKE_CASE : Tuple = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class lowercase_ ( __snake_case ): _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_INIT_CONFIGURATION _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = ConvBertTokenizer def __init__( self , lowercase_=None , lowercase_=None , lowercase_=True , lowercase_="[UNK]" , lowercase_="[SEP]" , lowercase_="[PAD]" , lowercase_="[CLS]" , lowercase_="[MASK]" , lowercase_=True , lowercase_=None , **lowercase_ , ): super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , tokenize_chinese_chars=lowercase_ , strip_accents=lowercase_ , **lowercase_ , ) _snake_case : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowercase_ ) != do_lower_case or normalizer_state.get("strip_accents" , lowercase_ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowercase_ ) != tokenize_chinese_chars ): _snake_case : Tuple = getattr(lowercase_ , normalizer_state.pop("type" ) ) _snake_case : List[str] = do_lower_case _snake_case : List[Any] = strip_accents _snake_case : List[Any] = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**lowercase_ ) _snake_case : Optional[Any] = do_lower_case def UpperCamelCase ( self , lowercase_ , lowercase_=None ): _snake_case : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): _snake_case : str = [self.sep_token_id] _snake_case : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): _snake_case : int = self._tokenizer.model.save(lowercase_ , name=lowercase_ ) return tuple(lowercase_ )
670
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
1
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=5 ) -> Union[str, Any]: '''simple docstring''' assert masked_input.count("<mask>" ) == 1 _snake_case : Dict = torch.tensor(tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) ).unsqueeze(0 ) # Batch size 1 _snake_case : Any = model(__lowercase )[0] # The last hidden-state is the first element of the output tuple _snake_case : Dict = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() _snake_case : Tuple = logits[0, masked_index, :] _snake_case : List[str] = logits.softmax(dim=0 ) _snake_case ,_snake_case : Optional[Any] = prob.topk(k=__lowercase , dim=0 ) _snake_case : Optional[int] = " ".join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(__lowercase ) )] ) _snake_case : List[str] = tokenizer.mask_token _snake_case : Dict = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(" " ) ): _snake_case : int = predicted_token_bpe.replace("\u2581" , " " ) if " {0}".format(__lowercase ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(" {0}".format(__lowercase ) , __lowercase ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(__lowercase , __lowercase ), values[index].item(), predicted_token, ) ) return topk_filled_outputs __SCREAMING_SNAKE_CASE : Union[str, Any] = CamembertTokenizer.from_pretrained('camembert-base') __SCREAMING_SNAKE_CASE : List[Any] = CamembertForMaskedLM.from_pretrained('camembert-base') model.eval() __SCREAMING_SNAKE_CASE : Optional[Any] = 'Le camembert est <mask> :)' print(fill_mask(masked_input, model, tokenizer, topk=3))
670
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : int = { 'configuration_distilbert': [ 'DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DistilBertConfig', 'DistilBertOnnxConfig', ], 'tokenization_distilbert': ['DistilBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Optional[int] = ['DistilBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ 'DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DistilBertForMaskedLM', 'DistilBertForMultipleChoice', 'DistilBertForQuestionAnswering', 'DistilBertForSequenceClassification', 'DistilBertForTokenClassification', 'DistilBertModel', 'DistilBertPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Optional[Any] = [ 'TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDistilBertForMaskedLM', 'TFDistilBertForMultipleChoice', 'TFDistilBertForQuestionAnswering', 'TFDistilBertForSequenceClassification', 'TFDistilBertForTokenClassification', 'TFDistilBertMainLayer', 'TFDistilBertModel', 'TFDistilBertPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ 'FlaxDistilBertForMaskedLM', 'FlaxDistilBertForMultipleChoice', 'FlaxDistilBertForQuestionAnswering', 'FlaxDistilBertForSequenceClassification', 'FlaxDistilBertForTokenClassification', 'FlaxDistilBertModel', 'FlaxDistilBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : Tuple = Rectangle(height=0.5 , width=0.5 ) _snake_case : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : List[str] = [mem.copy() for i in range(6 )] _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("CPU" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : int = [mem.copy() for i in range(4 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = Text("GPU" , font_size=24 ) _snake_case : Optional[int] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = Text("Model" , font_size=24 ) _snake_case : Dict = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case : Union[str, Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : Union[str, Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Union[str, Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Optional[Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _snake_case : List[Any] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : int = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Dict = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) _snake_case : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
670
1
import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel __SCREAMING_SNAKE_CASE : List[str] = { 'gwf-440k': { 'url': 'https://model-server.zqevans2.workers.dev/gwf-440k.ckpt', 'sample_rate': 4_8_0_0_0, 'sample_size': 6_5_5_3_6, }, 'jmann-small-190k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt', 'sample_rate': 4_8_0_0_0, 'sample_size': 6_5_5_3_6, }, 'jmann-large-580k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt', 'sample_rate': 4_8_0_0_0, 'sample_size': 1_3_1_0_7_2, }, 'maestro-uncond-150k': { 'url': 'https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt', 'sample_rate': 1_6_0_0_0, 'sample_size': 6_5_5_3_6, }, 'unlocked-uncond-250k': { 'url': 'https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt', 'sample_rate': 1_6_0_0_0, 'sample_size': 6_5_5_3_6, }, 'honk-140k': { 'url': 'https://model-server.zqevans2.workers.dev/honk-140k.ckpt', 'sample_rate': 1_6_0_0_0, 'sample_size': 6_5_5_3_6, }, } def snake_case (__lowercase , __lowercase ) -> Any: '''simple docstring''' return torch.atana(__lowercase , __lowercase ) / math.pi * 2 def snake_case (__lowercase ) -> List[Any]: '''simple docstring''' _snake_case : Dict = torch.sin(t * math.pi / 2 ) ** 2 _snake_case : List[Any] = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): pass class lowercase_ ( nn.Module ): def __init__( self , lowercase_ ): super().__init__() _snake_case : Union[str, Any] = DiffusionAttnUnetaD(lowercase_ , n_attn_layers=4 ) _snake_case : str = deepcopy(self.diffusion ) _snake_case : str = torch.quasirandom.SobolEngine(1 , scramble=lowercase_ ) def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' _snake_case : Union[str, Any] = MODELS_MAP[model_name]["url"] os.system(F"""wget {url} ./""" ) return F"""./{model_name}.ckpt""" __SCREAMING_SNAKE_CASE : Optional[int] = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', } __SCREAMING_SNAKE_CASE : Optional[int] = { '8': 'resnets.0', '9': 'attentions.0', '10': 'resnets.1', '11': 'attentions.1', '12': 'resnets.2', '13': 'attentions.2', } __SCREAMING_SNAKE_CASE : int = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', '8': 'resnets.3', '9': 'attentions.3', '10': 'resnets.4', '11': 'attentions.4', '12': 'resnets.5', '13': 'attentions.5', } __SCREAMING_SNAKE_CASE : List[str] = { '0': 'resnets.0', '1': 'resnets.1', '2': 'resnets.2', '4': 'resnets.0', '5': 'resnets.1', '6': 'resnets.2', } __SCREAMING_SNAKE_CASE : Optional[int] = { 'skip': 'conv_skip', 'main.0': 'conv_1', 'main.1': 'group_norm_1', 'main.3': 'conv_2', 'main.4': 'group_norm_2', } __SCREAMING_SNAKE_CASE : Tuple = { 'norm': 'group_norm', 'qkv_proj': ['query', 'key', 'value'], 'out_proj': ['proj_attn'], } def snake_case (__lowercase ) -> str: '''simple docstring''' if name.startswith("skip" ): return name.replace("skip" , RES_CONV_MAP["skip"] ) # name has to be of format main.{digit} if not name.startswith("main." ): raise ValueError(F"""ResConvBlock error with {name}""" ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def snake_case (__lowercase ) -> Union[str, Any]: '''simple docstring''' for key, value in ATTN_MAP.items(): if name.startswith(__lowercase ) and not isinstance(__lowercase , __lowercase ): return name.replace(__lowercase , __lowercase ) elif name.startswith(__lowercase ): return [name.replace(__lowercase , __lowercase ) for v in value] raise ValueError(F"""Attn error with {name}""" ) def snake_case (__lowercase , __lowercase=13 ) -> Optional[int]: '''simple docstring''' _snake_case : Tuple = input_string if string.split("." )[0] == "timestep_embed": return string.replace("timestep_embed" , "time_proj" ) _snake_case : Tuple = 0 if string.startswith("net.3." ): depth += 1 _snake_case : Optional[int] = string[6:] elif string.startswith("net." ): _snake_case : Any = string[4:] while string.startswith("main.7." ): depth += 1 _snake_case : List[Any] = string[7:] if string.startswith("main." ): _snake_case : Optional[int] = string[5:] # mid block if string[:2].isdigit(): _snake_case : List[Any] = string[:2] _snake_case : Optional[int] = string[2:] else: _snake_case : Optional[int] = string[0] _snake_case : Optional[int] = string[1:] if depth == max_depth: _snake_case : Optional[Any] = MID_NUM_TO_LAYER[layer_num] _snake_case : Tuple = "mid_block" elif depth > 0 and int(__lowercase ) < 7: _snake_case : Any = DOWN_NUM_TO_LAYER[layer_num] _snake_case : List[str] = F"""down_blocks.{depth}""" elif depth > 0 and int(__lowercase ) > 7: _snake_case : Tuple = UP_NUM_TO_LAYER[layer_num] _snake_case : Union[str, Any] = F"""up_blocks.{max_depth - depth - 1}""" elif depth == 0: _snake_case : Optional[int] = DEPTH_0_TO_LAYER[layer_num] _snake_case : List[Any] = F"""up_blocks.{max_depth - 1}""" if int(__lowercase ) > 3 else "down_blocks.0" if not string_left.startswith("." ): raise ValueError(F"""Naming error with {input_string} and string_left: {string_left}.""" ) _snake_case : Union[str, Any] = string_left[1:] if "resnets" in new_layer: _snake_case : List[str] = convert_resconv_naming(__lowercase ) elif "attentions" in new_layer: _snake_case : Tuple = convert_attn_naming(__lowercase ) _snake_case : Any = new_string_left if not isinstance(__lowercase , __lowercase ): _snake_case : Optional[int] = prefix + "." + new_layer + "." + string_left else: _snake_case : int = [prefix + "." + new_layer + "." + s for s in string_left] return new_string def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Tuple = {} for k, v in state_dict.items(): if k.endswith("kernel" ): # up- and downsample layers, don't have trainable weights continue _snake_case : Dict = rename(__lowercase ) # check if we need to transform from Conv => Linear for attention if isinstance(__lowercase , __lowercase ): _snake_case : Tuple = transform_conv_attns(__lowercase , __lowercase , __lowercase ) else: _snake_case : str = v return new_state_dict def snake_case (__lowercase , __lowercase , __lowercase ) -> Optional[Any]: '''simple docstring''' if len(__lowercase ) == 1: if len(v.shape ) == 3: # weight _snake_case : Tuple = v[:, :, 0] else: # bias _snake_case : str = v else: # qkv matrices _snake_case : Any = v.shape[0] _snake_case : Union[str, Any] = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: _snake_case : Tuple = v[i * single_shape : (i + 1) * single_shape, :, 0] else: _snake_case : List[Any] = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case : List[str] = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) _snake_case : Optional[int] = args.model_path.split("/" )[-1].split("." )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), F"""Make sure to provide one of the official model names {MODELS_MAP.keys()}""" _snake_case : Tuple = download(__lowercase ) _snake_case : Any = MODELS_MAP[model_name]["sample_rate"] _snake_case : Any = MODELS_MAP[model_name]["sample_size"] _snake_case : Union[str, Any] = Object() _snake_case : Union[str, Any] = sample_size _snake_case : Union[str, Any] = sample_rate _snake_case : Dict = 0 _snake_case : Optional[Any] = UNetaDModel(sample_size=__lowercase , sample_rate=__lowercase ) _snake_case : List[Any] = diffusers_model.state_dict() _snake_case : Any = DiffusionUncond(__lowercase ) orig_model.load_state_dict(torch.load(args.model_path , map_location=__lowercase )["state_dict"] ) _snake_case : Dict = orig_model.diffusion_ema.eval() _snake_case : Dict = orig_model.state_dict() _snake_case : List[str] = rename_orig_weights(__lowercase ) _snake_case : Union[str, Any] = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) _snake_case : Union[str, Any] = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(__lowercase ) == 0, F"""Problem with {renamed_minus_diffusers}""" assert all(k.endswith("kernel" ) for k in list(__lowercase ) ), F"""Problem with {diffusers_minus_renamed}""" for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), F"""Shape for {key} doesn't match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}""" if key == "time_proj.weight": _snake_case : int = value.squeeze() _snake_case : List[str] = value diffusers_model.load_state_dict(__lowercase ) _snake_case : Optional[Any] = 100 _snake_case : Union[str, Any] = 33 _snake_case : Optional[int] = IPNDMScheduler(num_train_timesteps=__lowercase ) _snake_case : Optional[int] = torch.manual_seed(__lowercase ) _snake_case : Optional[Any] = torch.randn([1, 2, config.sample_size] , generator=__lowercase ).to(__lowercase ) _snake_case : Optional[Any] = torch.linspace(1 , 0 , steps + 1 , device=__lowercase )[:-1] _snake_case : int = get_crash_schedule(__lowercase ) _snake_case : List[str] = DanceDiffusionPipeline(unet=__lowercase , scheduler=__lowercase ) _snake_case : str = torch.manual_seed(33 ) _snake_case : List[Any] = pipe(num_inference_steps=__lowercase , generator=__lowercase ).audios _snake_case : str = sampling.iplms_sample(__lowercase , __lowercase , __lowercase , {} ) _snake_case : Optional[int] = generated.clamp(-1 , 1 ) _snake_case : str = (generated - audio).abs().sum() _snake_case : int = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print("Diff sum" , __lowercase ) print("Diff max" , __lowercase ) assert diff_max < 1e-3, F"""Diff max: {diff_max} is too much :-/""" print(F"""Conversion for {model_name} successful!""" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') __SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() main(args)
670
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
1
import math def snake_case () -> None: '''simple docstring''' _snake_case : int = input("Enter message: " ) _snake_case : Any = int(input(F"""Enter key [2-{len(__lowercase ) - 1}]: """ ) ) _snake_case : List[Any] = input("Encryption/Decryption [e/d]: " ) if mode.lower().startswith("e" ): _snake_case : Optional[int] = encrypt_message(__lowercase , __lowercase ) elif mode.lower().startswith("d" ): _snake_case : Tuple = decrypt_message(__lowercase , __lowercase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F"""Output:\n{text + '|'}""" ) def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [""] * key for col in range(__lowercase ): _snake_case : Tuple = col while pointer < len(__lowercase ): cipher_text[col] += message[pointer] pointer += key return "".join(__lowercase ) def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[Any] = math.ceil(len(__lowercase ) / key ) _snake_case : str = key _snake_case : Union[str, Any] = (num_cols * num_rows) - len(__lowercase ) _snake_case : Tuple = [""] * num_cols _snake_case : List[str] = 0 _snake_case : str = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _snake_case : Union[str, Any] = 0 row += 1 return "".join(__lowercase ) if __name__ == "__main__": import doctest doctest.testmod() main()
670
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowercase_ ( __snake_case ): _lowerCamelCase = 'roc_bert' def __init__( self , lowercase_=30_522 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3_072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1e-12 , lowercase_=True , lowercase_=0 , lowercase_="absolute" , lowercase_=None , lowercase_=True , lowercase_=True , lowercase_=768 , lowercase_=910 , lowercase_=512 , lowercase_=24_858 , lowercase_=True , **lowercase_ , ): _snake_case : int = vocab_size _snake_case : Union[str, Any] = max_position_embeddings _snake_case : Union[str, Any] = hidden_size _snake_case : Dict = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : List[Any] = hidden_act _snake_case : Optional[int] = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : List[Any] = type_vocab_size _snake_case : int = layer_norm_eps _snake_case : Optional[Any] = use_cache _snake_case : List[Any] = enable_pronunciation _snake_case : Dict = enable_shape _snake_case : Dict = pronunciation_embed_dim _snake_case : Tuple = pronunciation_vocab_size _snake_case : Tuple = shape_embed_dim _snake_case : List[str] = shape_vocab_size _snake_case : Dict = concat_input _snake_case : int = position_embedding_type _snake_case : int = classifier_dropout super().__init__(pad_token_id=lowercase_ , **lowercase_ )
670
1
def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : int = len(__lowercase ) _snake_case : int = len(__lowercase ) _snake_case : int = ( first_str_length if first_str_length > second_str_length else second_str_length ) _snake_case : list = [] for char_count in range(__lowercase ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(__lowercase ) if __name__ == "__main__": print(alternative_string_arrange('AB', 'XYZ'), end=' ')
670
from cva import destroyAllWindows, imread, imshow, waitKey def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case ,_snake_case : int = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(__lowercase ): for j in range(__lowercase ): _snake_case : Optional[Any] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image __SCREAMING_SNAKE_CASE : Optional[Any] = imread('image_data/lena.jpg', 1) # convert to its negative __SCREAMING_SNAKE_CASE : Tuple = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
670
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : int = {'configuration_focalnet': ['FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FocalNetConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ 'FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'FocalNetForImageClassification', 'FocalNetForMaskedImageModeling', 'FocalNetBackbone', 'FocalNetModel', 'FocalNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __SCREAMING_SNAKE_CASE : List[str] = Mapping[str, np.ndarray] __SCREAMING_SNAKE_CASE : List[Any] = Mapping[str, Any] # Is a nested dict. __SCREAMING_SNAKE_CASE : List[Any] = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class lowercase_ : _lowerCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _lowerCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _lowerCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _lowerCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _lowerCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _lowerCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _lowerCamelCase = None # Templates used to generate this protein (prediction-only) _lowerCamelCase = None # Chain corresponding to each parent _lowerCamelCase = None def snake_case (__lowercase ) -> Protein: '''simple docstring''' _snake_case : str = r"(\[[A-Z]+\]\n)" _snake_case : List[str] = [tag.strip() for tag in re.split(__lowercase , __lowercase ) if len(__lowercase ) > 0] _snake_case : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("\n" ) for l in tags[1::2]] ) _snake_case : List[str] = ["N", "CA", "C"] _snake_case : Any = None _snake_case : Union[str, Any] = None _snake_case : Optional[int] = None for g in groups: if "[PRIMARY]" == g[0]: _snake_case : Tuple = g[1][0].strip() for i in range(len(__lowercase ) ): if seq[i] not in residue_constants.restypes: _snake_case : Tuple = "X" # FIXME: strings are immutable _snake_case : int = np.array( [residue_constants.restype_order.get(__lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: _snake_case : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(__lowercase , g[1][axis].split() ) ) ) _snake_case : Dict = np.array(__lowercase ) _snake_case : Dict = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: _snake_case : int = np.array(list(map({"-": 0, "+": 1}.get , g[1][0].strip() ) ) ) _snake_case : Any = np.zeros( ( len(__lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : Dict = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__lowercase , atom_mask=__lowercase , aatype=__lowercase , residue_index=np.arange(len(__lowercase ) ) , b_factors=__lowercase , ) def snake_case (__lowercase , __lowercase = 0 ) -> List[str]: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[Any] = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) _snake_case : str = prot.parents _snake_case : str = prot.parents_chain_index if parents is not None and parents_chain_index is not None: _snake_case : int = [p for i, p in zip(__lowercase , __lowercase ) if i == chain_id] if parents is None or len(__lowercase ) == 0: _snake_case : Optional[int] = ["N/A"] pdb_headers.append(F"""PARENT {' '.join(__lowercase )}""" ) return pdb_headers def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[int] = pdb_str.split("\n" ) _snake_case : List[str] = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) _snake_case : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: _snake_case : str = [] if prot.parents_chain_index is not None: _snake_case : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(__lowercase ) , [] ) parent_dict[str(__lowercase )].append(__lowercase ) _snake_case : Any = max([int(__lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): _snake_case : Tuple = parent_dict.get(str(__lowercase ) , ["N/A"] ) parents_per_chain.append(__lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: _snake_case : List[str] = [["N/A"]] def make_parent_line(__lowercase ) -> str: return F"""PARENT {' '.join(__lowercase )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) _snake_case : int = 0 for i, l in enumerate(__lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__lowercase ): _snake_case : Tuple = parents_per_chain[chain_counter] else: _snake_case : str = ["N/A"] out_pdb_lines.append(make_parent_line(__lowercase ) ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = residue_constants.restypes + ["X"] def res_atoa(__lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , "UNK" ) _snake_case : Optional[int] = residue_constants.atom_types _snake_case : List[str] = [] _snake_case : Tuple = prot.atom_mask _snake_case : List[str] = prot.aatype _snake_case : int = prot.atom_positions _snake_case : int = prot.residue_index.astype(np.intaa ) _snake_case : List[Any] = prot.b_factors _snake_case : str = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("Invalid aatypes." ) _snake_case : Union[str, Any] = get_pdb_headers(__lowercase ) if len(__lowercase ) > 0: pdb_lines.extend(__lowercase ) _snake_case : Optional[Any] = aatype.shape[0] _snake_case : str = 1 _snake_case : Tuple = 0 _snake_case : int = string.ascii_uppercase _snake_case : Optional[Any] = None # Add all atom sites. for i in range(__lowercase ): _snake_case : Dict = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue _snake_case : List[Any] = "ATOM" _snake_case : Union[str, Any] = atom_name if len(__lowercase ) == 4 else F""" {atom_name}""" _snake_case : str = "" _snake_case : str = "" _snake_case : Any = 1.00 _snake_case : str = atom_name[0] # Protein supports only C, N, O, S, this works. _snake_case : Dict = "" _snake_case : Any = "A" if chain_index is not None: _snake_case : List[Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! _snake_case : Optional[int] = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(__lowercase ) atom_index += 1 _snake_case : Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: _snake_case : Optional[int] = True _snake_case : Union[str, Any] = chain_index[i + 1] if should_terminate: # Close the chain. _snake_case : List[str] = "TER" _snake_case : str = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(__lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__lowercase , __lowercase ) ) pdb_lines.append("END" ) pdb_lines.append("" ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Protein: '''simple docstring''' return Protein( aatype=features["aatype"] , atom_positions=result["final_atom_positions"] , atom_mask=result["final_atom_mask"] , residue_index=features["residue_index"] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["final_atom_mask"] ) , chain_index=__lowercase , remark=__lowercase , parents=__lowercase , parents_chain_index=__lowercase , )
670
1
def snake_case (__lowercase , __lowercase ) -> int: '''simple docstring''' while a != 0: _snake_case ,_snake_case : Dict = b % a, a return b def snake_case (__lowercase , __lowercase ) -> int: '''simple docstring''' if gcd(__lowercase , __lowercase ) != 1: _snake_case : List[Any] = F"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(__lowercase ) _snake_case ,_snake_case ,_snake_case : Tuple = 1, 0, a _snake_case ,_snake_case ,_snake_case : Union[str, Any] = 0, 1, m while va != 0: _snake_case : Union[str, Any] = ua // va _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : Optional[int] = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
670
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor'] _lowerCamelCase = 'SamImageProcessor' def __init__( self , lowercase_ ): super().__init__(lowercase_ ) _snake_case : Optional[Any] = self.image_processor _snake_case : Tuple = -10 _snake_case : str = self.image_processor.size["longest_edge"] def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_ = None , **lowercase_ , ): _snake_case : List[Any] = self.image_processor( lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # pop arguments that are not used in the foward but used nevertheless _snake_case : Any = encoding_image_processor["original_sizes"] if hasattr(lowercase_ , "numpy" ): # Checks if Torch or TF tensor _snake_case : int = original_sizes.numpy() _snake_case ,_snake_case ,_snake_case : Union[str, Any] = self._check_and_preprocess_points( input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , ) _snake_case : Dict = self._normalize_and_convert( lowercase_ , lowercase_ , input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , return_tensors=lowercase_ , ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_="pt" , ): if input_points is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : int = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] ) for point in input_points ] else: _snake_case : Dict = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ ) for point, original_size in zip(lowercase_ , lowercase_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _snake_case ,_snake_case : int = self._pad_points_and_labels(lowercase_ , lowercase_ ) _snake_case : Any = np.array(lowercase_ ) if input_labels is not None: _snake_case : Optional[Any] = np.array(lowercase_ ) if input_boxes is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : Optional[Any] = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] , is_bounding_box=lowercase_ ) for box in input_boxes ] else: _snake_case : List[str] = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ , is_bounding_box=lowercase_ ) for box, original_size in zip(lowercase_ , lowercase_ ) ] _snake_case : Tuple = np.array(lowercase_ ) if input_boxes is not None: if return_tensors == "pt": _snake_case : List[str] = torch.from_numpy(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[int] = tf.expand_dims(lowercase_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": _snake_case : Tuple = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : int = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : Tuple = tf.expand_dims(lowercase_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": _snake_case : Dict = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : str = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _snake_case : Optional[Any] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : List[Any] = tf.expand_dims(lowercase_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : List[Any] = max([point.shape[0] for point in input_points] ) _snake_case : List[str] = [] for i, point in enumerate(lowercase_ ): if point.shape[0] != expected_nb_points: _snake_case : Optional[Any] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _snake_case : Union[str, Any] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowercase_ ) _snake_case : Optional[Any] = processed_input_points return input_points, input_labels def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=False ): _snake_case ,_snake_case : Optional[int] = original_size _snake_case ,_snake_case : List[str] = self.image_processor._get_preprocess_shape(lowercase_ , longest_edge=lowercase_ ) _snake_case : Optional[Any] = deepcopy(lowercase_ ).astype(lowercase_ ) if is_bounding_box: _snake_case : str = coords.reshape(-1 , 2 , 2 ) _snake_case : Optional[Any] = coords[..., 0] * (new_w / old_w) _snake_case : Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: _snake_case : Optional[Any] = coords.reshape(-1 , 4 ) return coords def UpperCamelCase ( self , lowercase_=None , lowercase_=None , lowercase_=None , ): if input_points is not None: if hasattr(lowercase_ , "numpy" ): # Checks for TF or Torch tensor _snake_case : Union[str, Any] = input_points.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_points[0] , lowercase_ ): raise ValueError("Input points must be a list of list of floating points." ) _snake_case : Any = [np.array(lowercase_ ) for input_point in input_points] else: _snake_case : Optional[int] = None if input_labels is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : Tuple = input_labels.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_labels[0] , lowercase_ ): raise ValueError("Input labels must be a list of list integers." ) _snake_case : Tuple = [np.array(lowercase_ ) for label in input_labels] else: _snake_case : Optional[Any] = None if input_boxes is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : List[str] = input_boxes.numpy().tolist() if ( not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_boxes[0] , lowercase_ ) or not isinstance(input_boxes[0][0] , lowercase_ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) _snake_case : List[Any] = [np.array(lowercase_ ).astype(np.floataa ) for box in input_boxes] else: _snake_case : Optional[int] = None return input_points, input_labels, input_boxes @property def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(lowercase_ ) ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.image_processor.post_process_masks(*lowercase_ , **lowercase_ )
670
1
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 __SCREAMING_SNAKE_CASE : int = data_utils.TransfoXLTokenizer __SCREAMING_SNAKE_CASE : Dict = data_utils.TransfoXLCorpus __SCREAMING_SNAKE_CASE : List[Any] = data_utils __SCREAMING_SNAKE_CASE : List[str] = data_utils def snake_case (__lowercase , __lowercase , __lowercase , __lowercase ) -> Union[str, Any]: '''simple docstring''' if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(__lowercase , "rb" ) as fp: _snake_case : Optional[int] = pickle.load(__lowercase , encoding="latin1" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) _snake_case : List[Any] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["pretrained_vocab_file"] print(F"""Save vocabulary to {pytorch_vocab_dump_path}""" ) _snake_case : Optional[Any] = corpus.vocab.__dict__ torch.save(__lowercase , __lowercase ) _snake_case : List[str] = corpus.__dict__ corpus_dict_no_vocab.pop("vocab" , __lowercase ) _snake_case : Optional[Any] = pytorch_dump_folder_path + "/" + CORPUS_NAME print(F"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(__lowercase , __lowercase ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model _snake_case : int = os.path.abspath(__lowercase ) _snake_case : List[str] = os.path.abspath(__lowercase ) print(F"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": _snake_case : Any = TransfoXLConfig() else: _snake_case : Dict = TransfoXLConfig.from_json_file(__lowercase ) print(F"""Building PyTorch model from configuration: {config}""" ) _snake_case : str = TransfoXLLMHeadModel(__lowercase ) _snake_case : Optional[Any] = load_tf_weights_in_transfo_xl(__lowercase , __lowercase , __lowercase ) # Save pytorch-model _snake_case : Tuple = os.path.join(__lowercase , __lowercase ) _snake_case : Any = os.path.join(__lowercase , __lowercase ) print(F"""Save PyTorch model to {os.path.abspath(__lowercase )}""" ) torch.save(model.state_dict() , __lowercase ) print(F"""Save configuration file to {os.path.abspath(__lowercase )}""" ) with open(__lowercase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) __SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
670
def snake_case (__lowercase ) -> int: '''simple docstring''' if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _snake_case : Union[str, Any] = grid[0] for row_n in range(1 , len(__lowercase ) ): _snake_case : Union[str, Any] = grid[row_n] _snake_case : List[Any] = fill_row(__lowercase , __lowercase ) _snake_case : List[Any] = grid[row_n] return grid[-1][-1] def snake_case (__lowercase , __lowercase ) -> list: '''simple docstring''' current_row[0] += row_above[0] for cell_n in range(1 , len(__lowercase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
670
1
def snake_case (__lowercase = 50 ) -> int: '''simple docstring''' _snake_case : Union[str, Any] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F'''{solution() = }''')
670
import random def snake_case (__lowercase , __lowercase ) -> tuple: '''simple docstring''' _snake_case ,_snake_case ,_snake_case : List[Any] = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def snake_case (__lowercase , __lowercase ) -> List[Any]: '''simple docstring''' if index >= len(__lowercase ) or index < 0: return None _snake_case : Any = items[random.randint(0 , len(__lowercase ) - 1 )] _snake_case : Tuple = 0 _snake_case ,_snake_case ,_snake_case : Tuple = _partition(__lowercase , __lowercase ) _snake_case : Tuple = len(__lowercase ) _snake_case : List[str] = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
670
1
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : List[Any] = Rectangle(height=0.5 , width=0.5 ) _snake_case : int = Rectangle(height=0.25 , width=0.25 ) _snake_case : int = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : int = [mem.copy() for i in range(6 )] _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : List[str] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : List[Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = Text("CPU" , font_size=24 ) _snake_case : List[Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : str = [mem.copy() for i in range(4 )] _snake_case : int = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : List[str] = Text("GPU" , font_size=24 ) _snake_case : Any = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Optional[int] = [mem.copy() for i in range(6 )] _snake_case : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[int] = Text("Model" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : List[Any] = [] _snake_case : List[str] = [] _snake_case : Tuple = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) _snake_case : Optional[Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) model_cpu_arr.append(lowercase_ ) self.add(*lowercase_ , *lowercase_ , *lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : int = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Any = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) checkpoint.move_to([3, 0.5, 0] ) self.add(lowercase_ ) _snake_case : str = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Tuple = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) ckpt_arr.append(lowercase_ ) _snake_case : Optional[int] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(lowercase_ ) self.add(*lowercase_ , *lowercase_ ) _snake_case : str = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Union[str, Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Optional[int] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(lowercase_ ) _snake_case : Tuple = MarkupText( f"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) _snake_case : Dict = [meta_mem.copy() for i in range(6 )] _snake_case : List[Any] = [meta_mem.copy() for i in range(6 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Disk" , font_size=24 ) _snake_case : List[str] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(lowercase_ , run_time=3 ) , Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : List[str] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(FadeOut(lowercase_ ) ) _snake_case : Any = MarkupText(f"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ , run_time=3 ) ) self.play( FadeOut(lowercase_ , lowercase_ , *lowercase_ , *lowercase_ ) , ) self.wait()
670
from math import pow, sqrt def snake_case (*__lowercase ) -> bool: '''simple docstring''' _snake_case : str = len(__lowercase ) > 0 and all(value > 0.0 for value in values ) return result def snake_case (__lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
670
1
import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_utils from tqdm import auto as tqdm_lib __SCREAMING_SNAKE_CASE : Any = threading.Lock() __SCREAMING_SNAKE_CASE : Optional[logging.Handler] = None __SCREAMING_SNAKE_CASE : Optional[Any] = { 'debug': logging.DEBUG, 'info': logging.INFO, 'warning': logging.WARNING, 'error': logging.ERROR, 'critical': logging.CRITICAL, } __SCREAMING_SNAKE_CASE : Tuple = logging.WARNING __SCREAMING_SNAKE_CASE : Optional[int] = True def snake_case () -> Any: '''simple docstring''' _snake_case : List[Any] = os.getenv("TRANSFORMERS_VERBOSITY" , __lowercase ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"""Unknown option TRANSFORMERS_VERBOSITY={env_level_str}, """ F"""has to be one of: { ', '.join(log_levels.keys() ) }""" ) return _default_log_level def snake_case () -> str: '''simple docstring''' return __name__.split("." )[0] def snake_case () -> logging.Logger: '''simple docstring''' return logging.getLogger(_get_library_name() ) def snake_case () -> None: '''simple docstring''' global _default_handler with _lock: if _default_handler: # This library has already configured the library root logger. return _snake_case : Any = logging.StreamHandler() # Set sys.stderr as stream. _snake_case : List[str] = sys.stderr.flush # Apply our default configuration to the library root logger. _snake_case : List[str] = _get_library_root_logger() library_root_logger.addHandler(_default_handler ) library_root_logger.setLevel(_get_default_logging_level() ) _snake_case : Optional[int] = False def snake_case () -> None: '''simple docstring''' global _default_handler with _lock: if not _default_handler: return _snake_case : Optional[Any] = _get_library_root_logger() library_root_logger.removeHandler(_default_handler ) library_root_logger.setLevel(logging.NOTSET ) _snake_case : Optional[Any] = None def snake_case () -> Union[str, Any]: '''simple docstring''' return log_levels def snake_case (__lowercase = None ) -> logging.Logger: '''simple docstring''' if name is None: _snake_case : Dict = _get_library_name() _configure_library_root_logger() return logging.getLogger(__lowercase ) def snake_case () -> int: '''simple docstring''' _configure_library_root_logger() return _get_library_root_logger().getEffectiveLevel() def snake_case (__lowercase ) -> None: '''simple docstring''' _configure_library_root_logger() _get_library_root_logger().setLevel(__lowercase ) def snake_case () -> Union[str, Any]: '''simple docstring''' return set_verbosity(__lowercase ) def snake_case () -> Tuple: '''simple docstring''' return set_verbosity(__lowercase ) def snake_case () -> str: '''simple docstring''' return set_verbosity(__lowercase ) def snake_case () -> Optional[Any]: '''simple docstring''' return set_verbosity(__lowercase ) def snake_case () -> None: '''simple docstring''' _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().removeHandler(_default_handler ) def snake_case () -> None: '''simple docstring''' _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().addHandler(_default_handler ) def snake_case (__lowercase ) -> None: '''simple docstring''' _configure_library_root_logger() assert handler is not None _get_library_root_logger().addHandler(__lowercase ) def snake_case (__lowercase ) -> None: '''simple docstring''' _configure_library_root_logger() assert handler is not None and handler not in _get_library_root_logger().handlers _get_library_root_logger().removeHandler(__lowercase ) def snake_case () -> None: '''simple docstring''' _configure_library_root_logger() _snake_case : str = False def snake_case () -> None: '''simple docstring''' _configure_library_root_logger() _snake_case : str = True def snake_case () -> None: '''simple docstring''' _snake_case : Optional[Any] = _get_library_root_logger().handlers for handler in handlers: _snake_case : int = logging.Formatter("[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s" ) handler.setFormatter(__lowercase ) def snake_case () -> None: '''simple docstring''' _snake_case : Any = _get_library_root_logger().handlers for handler in handlers: handler.setFormatter(__lowercase ) def snake_case (self , *__lowercase , **__lowercase ) -> Optional[int]: '''simple docstring''' _snake_case : Tuple = os.getenv("TRANSFORMERS_NO_ADVISORY_WARNINGS" , __lowercase ) if no_advisory_warnings: return self.warning(*__lowercase , **__lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = warning_advice @functools.lru_cache(__lowercase ) def snake_case (self , *__lowercase , **__lowercase ) -> int: '''simple docstring''' self.warning(*__lowercase , **__lowercase ) __SCREAMING_SNAKE_CASE : str = warning_once class lowercase_ : def __init__( self , *lowercase_ , **lowercase_ ): # pylint: disable=unused-argument _snake_case : Optional[int] = args[0] if args else None def __iter__( self ): return iter(self._iterator ) def __getattr__( self , lowercase_ ): def empty_fn(*lowercase_ , **lowercase_ ): # pylint: disable=unused-argument return return empty_fn def __enter__( self ): return self def __exit__( self , lowercase_ , lowercase_ , lowercase_ ): return class lowercase_ : def __call__( self , *lowercase_ , **lowercase_ ): if _tqdm_active: return tqdm_lib.tqdm(*lowercase_ , **lowercase_ ) else: return EmptyTqdm(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): _snake_case : Dict = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self ): if _tqdm_active: return tqdm_lib.tqdm.get_lock() __SCREAMING_SNAKE_CASE : int = _tqdm_cls() def snake_case () -> bool: '''simple docstring''' global _tqdm_active return bool(_tqdm_active ) def snake_case () -> Union[str, Any]: '''simple docstring''' global _tqdm_active _snake_case : Dict = True hf_hub_utils.enable_progress_bars() def snake_case () -> Any: '''simple docstring''' global _tqdm_active _snake_case : int = False hf_hub_utils.disable_progress_bars()
670
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
1
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> float: '''simple docstring''' if days_between_payments <= 0: raise ValueError("days_between_payments must be > 0" ) if daily_interest_rate < 0: raise ValueError("daily_interest_rate must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return principal * daily_interest_rate * days_between_payments def snake_case (__lowercase , __lowercase , __lowercase , ) -> float: '''simple docstring''' if number_of_compounding_periods <= 0: raise ValueError("number_of_compounding_periods must be > 0" ) if nominal_annual_interest_rate_percentage < 0: raise ValueError("nominal_annual_interest_rate_percentage must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def snake_case (__lowercase , __lowercase , __lowercase , ) -> float: '''simple docstring''' if number_of_years <= 0: raise ValueError("number_of_years must be > 0" ) if nominal_annual_percentage_rate < 0: raise ValueError("nominal_annual_percentage_rate must be >= 0" ) if principal <= 0: raise ValueError("principal must be > 0" ) return compound_interest( __lowercase , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
670
from __future__ import annotations from typing import TypedDict class lowercase_ ( __snake_case ): _lowerCamelCase = 42 _lowerCamelCase = 42 def snake_case (__lowercase ) -> list[str]: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(__lowercase ) )] def snake_case (__lowercase ) -> BWTTransformDict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) _snake_case : List[str] = all_rotations(__lowercase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation _snake_case : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__lowercase ), } return response def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: _snake_case : Union[str, Any] = int(__lowercase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(__lowercase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) _snake_case : Optional[Any] = [""] * len(__lowercase ) for _ in range(len(__lowercase ) ): for i in range(len(__lowercase ) ): _snake_case : Tuple = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = 'Provide a string that I will generate its BWT transform: ' __SCREAMING_SNAKE_CASE : Optional[Any] = input(entry_msg).strip() __SCREAMING_SNAKE_CASE : int = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result['bwt_string']}\'''' ) __SCREAMING_SNAKE_CASE : List[str] = reverse_bwt(result['bwt_string'], result['idx_original_string']) print( F'''Reversing Burrows Wheeler transform for entry \'{result['bwt_string']}\' ''' F'''we get original string \'{original_string}\'''' )
670
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( 'stable diffusion controlnet', '0.22.0', 'Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.', standard_warn=False, stacklevel=3, )
670
1
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[str] = OrderedDict( [ ('align', 'EfficientNetImageProcessor'), ('beit', 'BeitImageProcessor'), ('bit', 'BitImageProcessor'), ('blip', 'BlipImageProcessor'), ('blip-2', 'BlipImageProcessor'), ('bridgetower', 'BridgeTowerImageProcessor'), ('chinese_clip', 'ChineseCLIPImageProcessor'), ('clip', 'CLIPImageProcessor'), ('clipseg', 'ViTImageProcessor'), ('conditional_detr', 'ConditionalDetrImageProcessor'), ('convnext', 'ConvNextImageProcessor'), ('convnextv2', 'ConvNextImageProcessor'), ('cvt', 'ConvNextImageProcessor'), ('data2vec-vision', 'BeitImageProcessor'), ('deformable_detr', 'DeformableDetrImageProcessor'), ('deit', 'DeiTImageProcessor'), ('deta', 'DetaImageProcessor'), ('detr', 'DetrImageProcessor'), ('dinat', 'ViTImageProcessor'), ('donut-swin', 'DonutImageProcessor'), ('dpt', 'DPTImageProcessor'), ('efficientformer', 'EfficientFormerImageProcessor'), ('efficientnet', 'EfficientNetImageProcessor'), ('flava', 'FlavaImageProcessor'), ('focalnet', 'BitImageProcessor'), ('git', 'CLIPImageProcessor'), ('glpn', 'GLPNImageProcessor'), ('groupvit', 'CLIPImageProcessor'), ('imagegpt', 'ImageGPTImageProcessor'), ('instructblip', 'BlipImageProcessor'), ('layoutlmv2', 'LayoutLMv2ImageProcessor'), ('layoutlmv3', 'LayoutLMv3ImageProcessor'), ('levit', 'LevitImageProcessor'), ('mask2former', 'Mask2FormerImageProcessor'), ('maskformer', 'MaskFormerImageProcessor'), ('mgp-str', 'ViTImageProcessor'), ('mobilenet_v1', 'MobileNetV1ImageProcessor'), ('mobilenet_v2', 'MobileNetV2ImageProcessor'), ('mobilevit', 'MobileViTImageProcessor'), ('mobilevit', 'MobileViTImageProcessor'), ('mobilevitv2', 'MobileViTImageProcessor'), ('nat', 'ViTImageProcessor'), ('oneformer', 'OneFormerImageProcessor'), ('owlvit', 'OwlViTImageProcessor'), ('perceiver', 'PerceiverImageProcessor'), ('pix2struct', 'Pix2StructImageProcessor'), ('poolformer', 'PoolFormerImageProcessor'), ('regnet', 'ConvNextImageProcessor'), ('resnet', 'ConvNextImageProcessor'), ('sam', 'SamImageProcessor'), ('segformer', 'SegformerImageProcessor'), ('swiftformer', 'ViTImageProcessor'), ('swin', 'ViTImageProcessor'), ('swin2sr', 'Swin2SRImageProcessor'), ('swinv2', 'ViTImageProcessor'), ('table-transformer', 'DetrImageProcessor'), ('timesformer', 'VideoMAEImageProcessor'), ('tvlt', 'TvltImageProcessor'), ('upernet', 'SegformerImageProcessor'), ('van', 'ConvNextImageProcessor'), ('videomae', 'VideoMAEImageProcessor'), ('vilt', 'ViltImageProcessor'), ('vit', 'ViTImageProcessor'), ('vit_hybrid', 'ViTHybridImageProcessor'), ('vit_mae', 'ViTImageProcessor'), ('vit_msn', 'ViTImageProcessor'), ('xclip', 'CLIPImageProcessor'), ('yolos', 'YolosImageProcessor'), ] ) __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: _snake_case : List[str] = model_type_to_module_name(__lowercase ) _snake_case : Any = importlib.import_module(F""".{module_name}""" , "transformers.models" ) try: return getattr(__lowercase , __lowercase ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(__lowercase , "__name__" , __lowercase ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _snake_case : int = importlib.import_module("transformers" ) if hasattr(__lowercase , __lowercase ): return getattr(__lowercase , __lowercase ) return None def snake_case (__lowercase , __lowercase = None , __lowercase = False , __lowercase = False , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = False , **__lowercase , ) -> Any: '''simple docstring''' _snake_case : Optional[int] = get_file_from_repo( __lowercase , __lowercase , cache_dir=__lowercase , force_download=__lowercase , resume_download=__lowercase , proxies=__lowercase , use_auth_token=__lowercase , revision=__lowercase , local_files_only=__lowercase , ) if resolved_config_file is None: logger.info( "Could not locate the image processor configuration file, will try to use the model config instead." ) return {} with open(__lowercase , encoding="utf-8" ) as reader: return json.load(__lowercase ) class lowercase_ : def __init__( self ): raise EnvironmentError( "AutoImageProcessor is designed to be instantiated " "using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method." ) @classmethod @replace_list_option_in_docstrings(lowercase_ ) def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : Tuple = kwargs.pop("config" , lowercase_ ) _snake_case : List[str] = kwargs.pop("trust_remote_code" , lowercase_ ) _snake_case : Optional[int] = True _snake_case ,_snake_case : List[Any] = ImageProcessingMixin.get_image_processor_dict(lowercase_ , **lowercase_ ) _snake_case : List[Any] = config_dict.get("image_processor_type" , lowercase_ ) _snake_case : Tuple = None if "AutoImageProcessor" in config_dict.get("auto_map" , {} ): _snake_case : Any = config_dict["auto_map"]["AutoImageProcessor"] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: _snake_case : Optional[Any] = config_dict.pop("feature_extractor_type" , lowercase_ ) if feature_extractor_class is not None: logger.warning( "Could not find image processor class in the image processor config or the model config. Loading" " based on pattern matching with the model's feature extractor configuration." ) _snake_case : List[str] = feature_extractor_class.replace("FeatureExtractor" , "ImageProcessor" ) if "AutoFeatureExtractor" in config_dict.get("auto_map" , {} ): _snake_case : List[str] = config_dict["auto_map"]["AutoFeatureExtractor"] _snake_case : Union[str, Any] = feature_extractor_auto_map.replace("FeatureExtractor" , "ImageProcessor" ) logger.warning( "Could not find image processor auto map in the image processor config or the model config." " Loading based on pattern matching with the model's feature extractor configuration." ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(lowercase_ , lowercase_ ): _snake_case : Dict = AutoConfig.from_pretrained(lowercase_ , **lowercase_ ) # It could be in `config.image_processor_type`` _snake_case : Optional[int] = getattr(lowercase_ , "image_processor_type" , lowercase_ ) if hasattr(lowercase_ , "auto_map" ) and "AutoImageProcessor" in config.auto_map: _snake_case : List[str] = config.auto_map["AutoImageProcessor"] if image_processor_class is not None: _snake_case : List[Any] = image_processor_class_from_name(lowercase_ ) _snake_case : List[Any] = image_processor_auto_map is not None _snake_case : Optional[int] = image_processor_class is not None or type(lowercase_ ) in IMAGE_PROCESSOR_MAPPING _snake_case : Dict = resolve_trust_remote_code( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if has_remote_code and trust_remote_code: _snake_case : str = get_class_from_dynamic_module( lowercase_ , lowercase_ , **lowercase_ ) _snake_case : List[Any] = kwargs.pop("code_revision" , lowercase_ ) if os.path.isdir(lowercase_ ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(lowercase_ , **lowercase_ ) elif image_processor_class is not None: return image_processor_class.from_dict(lowercase_ , **lowercase_ ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(lowercase_ ) in IMAGE_PROCESSOR_MAPPING: _snake_case : Dict = IMAGE_PROCESSOR_MAPPING[type(lowercase_ )] return image_processor_class.from_dict(lowercase_ , **lowercase_ ) raise ValueError( f"""Unrecognized image processor in {pretrained_model_name_or_path}. Should have a """ f"""`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following """ f"""`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}""" ) @staticmethod def UpperCamelCase ( lowercase_ , lowercase_ ): IMAGE_PROCESSOR_MAPPING.register(lowercase_ , lowercase_ )
670
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase_ : _lowerCamelCase = LEDConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=4 , ): _snake_case : Optional[int] = parent _snake_case : str = batch_size _snake_case : int = seq_length _snake_case : Dict = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = vocab_size _snake_case : str = hidden_size _snake_case : int = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : int = intermediate_size _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : int = max_position_embeddings _snake_case : Union[str, Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Any = bos_token_id _snake_case : str = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after _snake_case : List[Any] = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests _snake_case : List[str] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def UpperCamelCase ( self ): _snake_case : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Any = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) _snake_case : Optional[Any] = prepare_led_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) _snake_case : int = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]] , axis=-1 , ) _snake_case : List[Any] = global_attention_mask return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = TFLEDModel(config=lowercase_ ).get_decoder() _snake_case : Optional[Any] = inputs_dict["input_ids"] _snake_case : Optional[int] = input_ids[:1, :] _snake_case : int = inputs_dict["attention_mask"][:1, :] _snake_case : int = 1 # first forward pass _snake_case : str = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case : Tuple = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case : List[str] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case : str = model(lowercase_ , attention_mask=lowercase_ )[0] _snake_case : List[str] = model(lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_ , lowercase_ , rtol=1e-3 ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> List[Any]: '''simple docstring''' if attention_mask is None: _snake_case : int = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () _lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFLEDModelTester(self ) _snake_case : List[Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = tf.zeros_like(inputs_dict["attention_mask"] ) _snake_case : Tuple = 2 _snake_case : Dict = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) _snake_case : Tuple = True _snake_case : Union[str, Any] = self.model_tester.seq_length _snake_case : Union[str, Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ ): _snake_case : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(lowercase_ ): _snake_case : int = [t.numpy() for t in outputs.encoder_attentions] _snake_case : Optional[int] = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: _snake_case : Union[str, Any] = True _snake_case : Dict = False _snake_case : Any = False _snake_case : Any = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) _snake_case : Tuple = len(lowercase_ ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: _snake_case : int = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Any = model_class(lowercase_ ) _snake_case : Optional[Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine _snake_case : Optional[int] = True _snake_case : Optional[int] = True _snake_case : List[Any] = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # TODO: Head-masking not yet implement pass def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' return tf.constant(__lowercase , dtype=tf.intaa ) __SCREAMING_SNAKE_CASE : List[Any] = 1E-4 @slow @require_tf class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Dict = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here _snake_case : Union[str, Any] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[int] = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Union[str, Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Optional[Any] = model(**lowercase_ )[0] _snake_case : str = (1, 1_024, 768) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[Any] = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 ) def UpperCamelCase ( self ): _snake_case : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here _snake_case : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : int = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Tuple = model(**lowercase_ )[0] _snake_case : Any = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[int] = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 , rtol=1e-3 )
670
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'ViltImageProcessor' _lowerCamelCase = ('BertTokenizer', 'BertTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Any = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) _snake_case : int = self.image_processor def __call__( self , lowercase_ , lowercase_ = None , lowercase_ = True , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = 0 , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = True , lowercase_ = None , **lowercase_ , ): _snake_case : int = self.tokenizer( text=lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , stride=lowercase_ , pad_to_multiple_of=lowercase_ , return_token_type_ids=lowercase_ , return_attention_mask=lowercase_ , return_overflowing_tokens=lowercase_ , return_special_tokens_mask=lowercase_ , return_offsets_mapping=lowercase_ , return_length=lowercase_ , verbose=lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # add pixel_values + pixel_mask _snake_case : Optional[Any] = self.image_processor(lowercase_ , return_tensors=lowercase_ ) encoding.update(lowercase_ ) return encoding def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : List[Any] = self.tokenizer.model_input_names _snake_case : Dict = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def UpperCamelCase ( self ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , lowercase_ , ) return self.image_processor_class @property def UpperCamelCase ( self ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , lowercase_ , ) return self.image_processor
670
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
1
from cva import destroyAllWindows, imread, imshow, waitKey def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case ,_snake_case : int = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(__lowercase ): for j in range(__lowercase ): _snake_case : Optional[Any] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image __SCREAMING_SNAKE_CASE : Optional[Any] = imread('image_data/lena.jpg', 1) # convert to its negative __SCREAMING_SNAKE_CASE : Tuple = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
670
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Any = tempfile.mkdtemp() # fmt: off _snake_case : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _snake_case : Dict = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] _snake_case : Optional[int] = {"unk_token": "<unk>"} _snake_case : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) _snake_case : Any = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } _snake_case : Optional[Any] = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Union[str, Any] = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Tuple = self.get_tokenizer() _snake_case : Any = self.get_rust_tokenizer() _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) _snake_case : Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) _snake_case : List[Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : List[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Tuple = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : int = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Optional[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : str = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : Dict = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : int = processor(text=lowercase_ ) _snake_case : str = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[Any] = "lower newer" _snake_case : int = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[str] = self.get_tokenizer() _snake_case : Union[str, Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Dict = self.prepare_image_inputs() _snake_case : List[Any] = processor(images=lowercase_ , visual_prompt=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[Any] = self.get_tokenizer() _snake_case : str = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Any = processor.batch_decode(lowercase_ ) _snake_case : Any = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ )
670
1
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'b0': efficientnet.EfficientNetBa, 'b1': efficientnet.EfficientNetBa, 'b2': efficientnet.EfficientNetBa, 'b3': efficientnet.EfficientNetBa, 'b4': efficientnet.EfficientNetBa, 'b5': efficientnet.EfficientNetBa, 'b6': efficientnet.EfficientNetBa, 'b7': efficientnet.EfficientNetBa, } __SCREAMING_SNAKE_CASE : Any = { 'b0': { 'hidden_dim': 1_2_8_0, 'width_coef': 1.0, 'depth_coef': 1.0, 'image_size': 2_2_4, 'dropout_rate': 0.2, 'dw_padding': [], }, 'b1': { 'hidden_dim': 1_2_8_0, 'width_coef': 1.0, 'depth_coef': 1.1, 'image_size': 2_4_0, 'dropout_rate': 0.2, 'dw_padding': [1_6], }, 'b2': { 'hidden_dim': 1_4_0_8, 'width_coef': 1.1, 'depth_coef': 1.2, 'image_size': 2_6_0, 'dropout_rate': 0.3, 'dw_padding': [5, 8, 1_6], }, 'b3': { 'hidden_dim': 1_5_3_6, 'width_coef': 1.2, 'depth_coef': 1.4, 'image_size': 3_0_0, 'dropout_rate': 0.3, 'dw_padding': [5, 1_8], }, 'b4': { 'hidden_dim': 1_7_9_2, 'width_coef': 1.4, 'depth_coef': 1.8, 'image_size': 3_8_0, 'dropout_rate': 0.4, 'dw_padding': [6], }, 'b5': { 'hidden_dim': 2_0_4_8, 'width_coef': 1.6, 'depth_coef': 2.2, 'image_size': 4_5_6, 'dropout_rate': 0.4, 'dw_padding': [1_3, 2_7], }, 'b6': { 'hidden_dim': 2_3_0_4, 'width_coef': 1.8, 'depth_coef': 2.6, 'image_size': 5_2_8, 'dropout_rate': 0.5, 'dw_padding': [3_1], }, 'b7': { 'hidden_dim': 2_5_6_0, 'width_coef': 2.0, 'depth_coef': 3.1, 'image_size': 6_0_0, 'dropout_rate': 0.5, 'dw_padding': [1_8], }, } def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : Optional[int] = EfficientNetConfig() _snake_case : Any = CONFIG_MAP[model_name]["hidden_dim"] _snake_case : Optional[Any] = CONFIG_MAP[model_name]["width_coef"] _snake_case : Any = CONFIG_MAP[model_name]["depth_coef"] _snake_case : str = CONFIG_MAP[model_name]["image_size"] _snake_case : Optional[int] = CONFIG_MAP[model_name]["dropout_rate"] _snake_case : Optional[int] = CONFIG_MAP[model_name]["dw_padding"] _snake_case : str = "huggingface/label-files" _snake_case : List[str] = "imagenet-1k-id2label.json" _snake_case : Any = 1_000 _snake_case : Any = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type="dataset" ) , "r" ) ) _snake_case : Any = {int(__lowercase ): v for k, v in idalabel.items()} _snake_case : Dict = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} return config def snake_case () -> Tuple: '''simple docstring''' _snake_case : Dict = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case : Tuple = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return im def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case : Optional[Any] = CONFIG_MAP[model_name]["image_size"] _snake_case : Union[str, Any] = EfficientNetImageProcessor( size={"height": size, "width": size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.47853944, 0.4732864, 0.47434163] , do_center_crop=__lowercase , ) return preprocessor def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case : List[Any] = [v.split("_" )[0].split("block" )[1] for v in original_param_names if v.startswith("block" )] _snake_case : Any = sorted(set(__lowercase ) ) _snake_case : Dict = len(__lowercase ) _snake_case : Tuple = {b: str(__lowercase ) for b, i in zip(__lowercase , range(__lowercase ) )} _snake_case : List[str] = [] rename_keys.append(("stem_conv/kernel:0", "embeddings.convolution.weight") ) rename_keys.append(("stem_bn/gamma:0", "embeddings.batchnorm.weight") ) rename_keys.append(("stem_bn/beta:0", "embeddings.batchnorm.bias") ) rename_keys.append(("stem_bn/moving_mean:0", "embeddings.batchnorm.running_mean") ) rename_keys.append(("stem_bn/moving_variance:0", "embeddings.batchnorm.running_var") ) for b in block_names: _snake_case : str = block_name_mapping[b] rename_keys.append((F"""block{b}_expand_conv/kernel:0""", F"""encoder.blocks.{hf_b}.expansion.expand_conv.weight""") ) rename_keys.append((F"""block{b}_expand_bn/gamma:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.weight""") ) rename_keys.append((F"""block{b}_expand_bn/beta:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.bias""") ) rename_keys.append( (F"""block{b}_expand_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.running_mean""") ) rename_keys.append( (F"""block{b}_expand_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.running_var""") ) rename_keys.append( (F"""block{b}_dwconv/depthwise_kernel:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight""") ) rename_keys.append((F"""block{b}_bn/gamma:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight""") ) rename_keys.append((F"""block{b}_bn/beta:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias""") ) rename_keys.append( (F"""block{b}_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean""") ) rename_keys.append( (F"""block{b}_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var""") ) rename_keys.append((F"""block{b}_se_reduce/kernel:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.reduce.weight""") ) rename_keys.append((F"""block{b}_se_reduce/bias:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.reduce.bias""") ) rename_keys.append((F"""block{b}_se_expand/kernel:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.expand.weight""") ) rename_keys.append((F"""block{b}_se_expand/bias:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.expand.bias""") ) rename_keys.append( (F"""block{b}_project_conv/kernel:0""", F"""encoder.blocks.{hf_b}.projection.project_conv.weight""") ) rename_keys.append((F"""block{b}_project_bn/gamma:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.weight""") ) rename_keys.append((F"""block{b}_project_bn/beta:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.bias""") ) rename_keys.append( (F"""block{b}_project_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.running_mean""") ) rename_keys.append( (F"""block{b}_project_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.running_var""") ) rename_keys.append(("top_conv/kernel:0", "encoder.top_conv.weight") ) rename_keys.append(("top_bn/gamma:0", "encoder.top_bn.weight") ) rename_keys.append(("top_bn/beta:0", "encoder.top_bn.bias") ) rename_keys.append(("top_bn/moving_mean:0", "encoder.top_bn.running_mean") ) rename_keys.append(("top_bn/moving_variance:0", "encoder.top_bn.running_var") ) _snake_case : Optional[Any] = {} for item in rename_keys: if item[0] in original_param_names: _snake_case : Any = "efficientnet." + item[1] _snake_case : List[Any] = "classifier.weight" _snake_case : str = "classifier.bias" return key_mapping def snake_case (__lowercase , __lowercase , __lowercase ) -> int: '''simple docstring''' for key, value in tf_params.items(): if "normalization" in key: continue _snake_case : Optional[int] = key_mapping[key] if "_conv" in key and "kernel" in key: _snake_case : Union[str, Any] = torch.from_numpy(__lowercase ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: _snake_case : Tuple = torch.from_numpy(__lowercase ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: _snake_case : Optional[int] = torch.from_numpy(np.transpose(__lowercase ) ) else: _snake_case : Union[str, Any] = torch.from_numpy(__lowercase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(__lowercase ) @torch.no_grad() def snake_case (__lowercase , __lowercase , __lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' _snake_case : List[str] = model_classes[model_name]( include_top=__lowercase , weights="imagenet" , input_tensor=__lowercase , input_shape=__lowercase , pooling=__lowercase , classes=1_000 , classifier_activation="softmax" , ) _snake_case : Union[str, Any] = original_model.trainable_variables _snake_case : Union[str, Any] = original_model.non_trainable_variables _snake_case : Union[str, Any] = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: _snake_case : Dict = param.numpy() _snake_case : int = list(tf_params.keys() ) # Load HuggingFace model _snake_case : Any = get_efficientnet_config(__lowercase ) _snake_case : int = EfficientNetForImageClassification(__lowercase ).eval() _snake_case : List[str] = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("Converting parameters..." ) _snake_case : Any = rename_keys(__lowercase ) replace_params(__lowercase , __lowercase , __lowercase ) # Initialize preprocessor and preprocess input image _snake_case : Dict = convert_image_processor(__lowercase ) _snake_case : Optional[int] = preprocessor(images=prepare_img() , return_tensors="pt" ) # HF model inference hf_model.eval() with torch.no_grad(): _snake_case : Union[str, Any] = hf_model(**__lowercase ) _snake_case : List[Any] = outputs.logits.detach().numpy() # Original model inference _snake_case : Tuple = False _snake_case : Optional[Any] = CONFIG_MAP[model_name]["image_size"] _snake_case : Optional[Any] = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) _snake_case : str = image.img_to_array(__lowercase ) _snake_case : Optional[int] = np.expand_dims(__lowercase , axis=0 ) _snake_case : List[Any] = original_model.predict(__lowercase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(__lowercase , __lowercase , atol=1e-3 ), "The predicted logits are not the same." print("Model outputs match!" ) if save_model: # Create folder to save model if not os.path.isdir(__lowercase ): os.mkdir(__lowercase ) # Save converted model and image processor hf_model.save_pretrained(__lowercase ) preprocessor.save_pretrained(__lowercase ) if push_to_hub: # Push model and image processor to hub print(F"""Pushing converted {model_name} to the hub...""" ) _snake_case : Optional[int] = F"""efficientnet-{model_name}""" preprocessor.push_to_hub(__lowercase ) hf_model.push_to_hub(__lowercase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='b0', type=str, help='Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].', ) parser.add_argument( '--pytorch_dump_folder_path', default='hf_model', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--save_model', action='store_true', help='Save model to local') parser.add_argument('--push_to_hub', action='store_true', help='Push model and image processor to the hub') __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
670
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
1
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ ): super().__init__() _snake_case : List[str] = nn.ModuleList(lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = True , ): for i, (image, scale, controlnet) in enumerate(zip(lowercase_ , lowercase_ , self.nets ) ): _snake_case ,_snake_case : Optional[int] = controlnet( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) # merge samples if i == 0: _snake_case ,_snake_case : Tuple = down_samples, mid_sample else: _snake_case : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase_ , lowercase_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , ): _snake_case : Tuple = 0 _snake_case : Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase_ , is_main_process=lowercase_ , save_function=lowercase_ , safe_serialization=lowercase_ , variant=lowercase_ , ) idx += 1 _snake_case : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : List[str] = 0 _snake_case : Optional[Any] = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case : Optional[Any] = pretrained_model_path while os.path.isdir(lowercase_ ): _snake_case : int = ControlNetModel.from_pretrained(lowercase_ , **lowercase_ ) controlnets.append(lowercase_ ) idx += 1 _snake_case : str = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(lowercase_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(lowercase_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(lowercase_ )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(lowercase_ )
670
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'asapp/sew-d-tiny-100k': 'https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class lowercase_ ( __snake_case ): _lowerCamelCase = 'sew-d' def __init__( self , lowercase_=32 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3_072 , lowercase_=2 , lowercase_=512 , lowercase_=256 , lowercase_=True , lowercase_=True , lowercase_=("p2c", "c2p") , lowercase_="layer_norm" , lowercase_="gelu_python" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.1 , lowercase_=0.02 , lowercase_=1e-7 , lowercase_=1e-5 , lowercase_="group" , lowercase_="gelu" , lowercase_=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase_=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase_=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase_=False , lowercase_=128 , lowercase_=16 , lowercase_=True , lowercase_=0.05 , lowercase_=10 , lowercase_=2 , lowercase_=0.0 , lowercase_=10 , lowercase_=0 , lowercase_="mean" , lowercase_=False , lowercase_=False , lowercase_=256 , lowercase_=0 , lowercase_=1 , lowercase_=2 , **lowercase_ , ): super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) _snake_case : Tuple = hidden_size _snake_case : Union[str, Any] = feat_extract_norm _snake_case : Optional[int] = feat_extract_activation _snake_case : str = list(lowercase_ ) _snake_case : Optional[Any] = list(lowercase_ ) _snake_case : Dict = list(lowercase_ ) _snake_case : Tuple = conv_bias _snake_case : Dict = num_conv_pos_embeddings _snake_case : Dict = num_conv_pos_embedding_groups _snake_case : Tuple = len(self.conv_dim ) _snake_case : Optional[int] = num_hidden_layers _snake_case : int = intermediate_size _snake_case : str = squeeze_factor _snake_case : Tuple = max_position_embeddings _snake_case : List[str] = position_buckets _snake_case : List[Any] = share_att_key _snake_case : int = relative_attention _snake_case : Optional[Any] = norm_rel_ebd _snake_case : List[Any] = list(lowercase_ ) _snake_case : int = hidden_act _snake_case : Optional[Any] = num_attention_heads _snake_case : List[str] = hidden_dropout _snake_case : Optional[Any] = attention_dropout _snake_case : int = activation_dropout _snake_case : List[Any] = feat_proj_dropout _snake_case : Union[str, Any] = final_dropout _snake_case : List[Any] = layer_norm_eps _snake_case : Optional[Any] = feature_layer_norm_eps _snake_case : Tuple = initializer_range _snake_case : List[str] = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," f"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" f"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _snake_case : Any = apply_spec_augment _snake_case : Union[str, Any] = mask_time_prob _snake_case : Dict = mask_time_length _snake_case : str = mask_time_min_masks _snake_case : Any = mask_feature_prob _snake_case : Optional[int] = mask_feature_length _snake_case : List[Any] = mask_feature_min_masks # ctc loss _snake_case : Any = ctc_loss_reduction _snake_case : List[str] = ctc_zero_infinity # sequence classification _snake_case : List[str] = use_weighted_layer_sum _snake_case : Tuple = classifier_proj_size @property def UpperCamelCase ( self ): return functools.reduce(operator.mul , self.conv_stride , 1 )
670
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'CLIPImageProcessor' _lowerCamelCase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Dict = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _snake_case : str = self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if images is not None: _snake_case : List[str] = self.image_processor(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if text is not None and images is not None: _snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase_ ) , tensor_type=lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
1
def snake_case (__lowercase ) -> list: '''simple docstring''' _snake_case : str = int(__lowercase ) if n_element < 1: _snake_case : List[str] = ValueError("a should be a positive number" ) raise my_error _snake_case : List[str] = [1] _snake_case ,_snake_case ,_snake_case : Optional[int] = (0, 0, 0) _snake_case : Tuple = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = input('Enter the last number (nth term) of the Hamming Number Series: ') print('Formula of Hamming Number Series => 2^i * 3^j * 5^k') __SCREAMING_SNAKE_CASE : Tuple = hamming(int(n)) print('-----------------------------------------------------') print(F'''The list with nth numbers is: {hamming_numbers}''') print('-----------------------------------------------------')
670
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
670
1
import os from math import logaa def snake_case (__lowercase = "base_exp.txt" ) -> int: '''simple docstring''' _snake_case : float = 0 _snake_case : Dict = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__lowercase ) , __lowercase ) ) ): _snake_case ,_snake_case : int = list(map(__lowercase , line.split("," ) ) ) if x * logaa(__lowercase ) > largest: _snake_case : Any = x * logaa(__lowercase ) _snake_case : List[Any] = i + 1 return result if __name__ == "__main__": print(solution())
670
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def snake_case (*__lowercase ) -> Dict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): _snake_case : Dict = list(__lowercase ) for i in range(len(__lowercase ) ): _snake_case : List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : str = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowercase , __lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def snake_case (__lowercase = None , __lowercase = 128 ) -> Any: '''simple docstring''' if function is None: return functools.partial(__lowercase , starting_batch_size=__lowercase ) _snake_case : List[str] = starting_batch_size def decorator(*__lowercase , **__lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _snake_case : Optional[Any] = list(inspect.signature(__lowercase ).parameters.keys() ) # Guard against user error if len(__lowercase ) < (len(__lowercase ) + 1): _snake_case : str = ", ".join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowercase , *__lowercase , **__lowercase ) except Exception as e: if should_reduce_batch_size(__lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
670
1
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __SCREAMING_SNAKE_CASE : int = 1_6 __SCREAMING_SNAKE_CASE : Tuple = 3_2 def snake_case (__lowercase , __lowercase = 16 ) -> List[Any]: '''simple docstring''' _snake_case : Union[str, Any] = AutoTokenizer.from_pretrained("bert-base-cased" ) _snake_case : Dict = load_dataset("glue" , "mrpc" ) def tokenize_function(__lowercase ): # max_length=None => use the model max length (it's actually the default) _snake_case : int = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=__lowercase , max_length=__lowercase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _snake_case : int = datasets.map( __lowercase , batched=__lowercase , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _snake_case : Optional[int] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(__lowercase ): # On TPU it's best to pad everything to the same length or training will be very slow. _snake_case : int = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _snake_case : Optional[Any] = 16 elif accelerator.mixed_precision != "no": _snake_case : List[str] = 8 else: _snake_case : str = None return tokenizer.pad( __lowercase , padding="longest" , max_length=__lowercase , pad_to_multiple_of=__lowercase , return_tensors="pt" , ) # Instantiate dataloaders. _snake_case : Optional[Any] = DataLoader( tokenized_datasets["train"] , shuffle=__lowercase , collate_fn=__lowercase , batch_size=__lowercase , drop_last=__lowercase ) _snake_case : Dict = DataLoader( tokenized_datasets["validation"] , shuffle=__lowercase , collate_fn=__lowercase , batch_size=__lowercase , drop_last=(accelerator.mixed_precision == "fp8") , ) return train_dataloader, eval_dataloader def snake_case (__lowercase , __lowercase ) -> Tuple: '''simple docstring''' _snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _snake_case : Optional[Any] = config["lr"] _snake_case : List[str] = int(config["num_epochs"] ) _snake_case : List[Any] = int(config["seed"] ) _snake_case : Union[str, Any] = int(config["batch_size"] ) _snake_case : str = evaluate.load("glue" , "mrpc" ) # If the batch size is too big we use gradient accumulation _snake_case : int = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _snake_case : int = batch_size // MAX_GPU_BATCH_SIZE _snake_case : int = MAX_GPU_BATCH_SIZE set_seed(__lowercase ) _snake_case ,_snake_case : Dict = get_dataloaders(__lowercase , __lowercase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _snake_case : str = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=__lowercase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer _snake_case : List[Any] = AdamW(params=model.parameters() , lr=__lowercase ) # Instantiate scheduler _snake_case : Any = get_linear_schedule_with_warmup( optimizer=__lowercase , num_warmup_steps=100 , num_training_steps=(len(__lowercase ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : Optional[Any] = accelerator.prepare( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Now we train the model for epoch in range(__lowercase ): model.train() for step, batch in enumerate(__lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _snake_case : Any = model(**__lowercase ) _snake_case : Tuple = outputs.loss _snake_case : str = loss / gradient_accumulation_steps accelerator.backward(__lowercase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _snake_case : Optional[int] = model(**__lowercase ) _snake_case : Dict = outputs.logits.argmax(dim=-1 ) _snake_case ,_snake_case : int = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=__lowercase , references=__lowercase , ) _snake_case : Optional[int] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , __lowercase ) def snake_case () -> Tuple: '''simple docstring''' _snake_case : Dict = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=__lowercase , default=__lowercase , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) _snake_case : Optional[Any] = parser.parse_args() _snake_case : Union[str, Any] = {"lr": 2e-5, "num_epochs": 3, "seed": 42, "batch_size": 16} training_function(__lowercase , __lowercase ) if __name__ == "__main__": main()
670
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
1
import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def snake_case (__lowercase , __lowercase=10 ) -> str: '''simple docstring''' _snake_case : List[str] = [] for _ in range(__lowercase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def snake_case (__lowercase , __lowercase=10 ) -> Optional[Any]: '''simple docstring''' _snake_case : Union[str, Any] = [] for step in range(__lowercase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Dict = os.path.join(__lowercase , "schedule.bin" ) torch.save(scheduler.state_dict() , __lowercase ) _snake_case : List[Any] = torch.load(__lowercase ) scheduler.load_state_dict(__lowercase ) return lrs @require_torch class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ ): self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for a, b in zip(lowercase_ , lowercase_ ): self.assertAlmostEqual(lowercase_ , lowercase_ , delta=lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[str] = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowercase_ ) _snake_case : Union[str, Any] = torch.tensor([0.4, 0.2, -0.5] ) _snake_case : Optional[int] = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _snake_case : List[str] = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(100 ): _snake_case : str = criterion(lowercase_ , lowercase_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def UpperCamelCase ( self ): _snake_case : str = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowercase_ ) _snake_case : Tuple = torch.tensor([0.4, 0.2, -0.5] ) _snake_case : Tuple = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _snake_case : List[Any] = Adafactor( params=[w] , lr=1e-2 , eps=(1e-30, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=lowercase_ , weight_decay=0.0 , relative_step=lowercase_ , scale_parameter=lowercase_ , warmup_init=lowercase_ , ) for _ in range(1_000 ): _snake_case : Tuple = criterion(lowercase_ , lowercase_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class lowercase_ ( unittest.TestCase ): _lowerCamelCase = nn.Linear(50 , 50 ) if is_torch_available() else None _lowerCamelCase = AdamW(m.parameters() , lr=10.0 ) if is_torch_available() else None _lowerCamelCase = 10 def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=None ): self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for a, b in zip(lowercase_ , lowercase_ ): self.assertAlmostEqual(lowercase_ , lowercase_ , delta=lowercase_ , msg=lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = {"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _snake_case : str = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1e-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): _snake_case ,_snake_case : Tuple = data _snake_case : Optional[Any] = scheduler_func(self.optimizer , **lowercase_ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _snake_case : Union[str, Any] = unwrap_schedule(lowercase_ , self.num_steps ) self.assertListAlmostEqual( lowercase_ , lowercase_ , tol=1e-2 , msg=f"""failed for {scheduler_func} in normal scheduler""" , ) _snake_case : List[str] = scheduler_func(self.optimizer , **lowercase_ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(lowercase_ ) # wrap to test picklability of the schedule _snake_case : List[str] = unwrap_and_save_reload_schedule(lowercase_ , self.num_steps ) self.assertListEqual(lowercase_ , lowercase_ , msg=f"""failed for {scheduler_func} in save and reload""" ) class lowercase_ : def __init__( self , lowercase_ ): _snake_case : Tuple = fn def __call__( self , *lowercase_ , **lowercase_ ): return self.fn(*lowercase_ , **lowercase_ ) @classmethod def UpperCamelCase ( self , lowercase_ ): _snake_case : str = list(map(self , scheduler.lr_lambdas ) )
670
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
1
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def snake_case (__lowercase="" ) -> str: '''simple docstring''' _snake_case : List[Any] = tempfile.mkdtemp() return os.path.join(__lowercase , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Tuple = torch.rand(12 , dtype=torch.floataa ) - 0.5 _snake_case : str = AgentAudio(lowercase_ ) _snake_case : Any = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase_ , agent_type.to_raw() , atol=1e-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowercase_ ) ) # Ensure that the file contains the same value as the original tensor _snake_case ,_snake_case : Optional[Any] = sf.read(lowercase_ ) self.assertTrue(torch.allclose(lowercase_ , torch.tensor(lowercase_ ) , atol=1e-4 ) ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = torch.rand(12 , dtype=torch.floataa ) - 0.5 _snake_case : Any = get_new_path(suffix=".wav" ) sf.write(lowercase_ , lowercase_ , 16_000 ) _snake_case : List[Any] = AgentAudio(lowercase_ ) self.assertTrue(torch.allclose(lowercase_ , agent_type.to_raw() , atol=1e-4 ) ) self.assertEqual(agent_type.to_string() , lowercase_ ) @require_vision @require_torch class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : str = torch.randint(0 , 256 , (64, 64, 3) ) _snake_case : Tuple = AgentImage(lowercase_ ) _snake_case : Dict = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase_ , agent_type._tensor , atol=1e-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : List[str] = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" _snake_case : Tuple = Image.open(lowercase_ ) _snake_case : Dict = AgentImage(lowercase_ ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" _snake_case : List[Any] = Image.open(lowercase_ ) _snake_case : Dict = AgentImage(lowercase_ ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Optional[Any] = "Hey!" _snake_case : Dict = AgentText(lowercase_ ) self.assertEqual(lowercase_ , agent_type.to_string() ) self.assertEqual(lowercase_ , agent_type.to_raw() ) self.assertEqual(lowercase_ , lowercase_ )
670
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : Tuple = Rectangle(height=0.5 , width=0.5 ) _snake_case : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : List[str] = [mem.copy() for i in range(6 )] _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("CPU" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : int = [mem.copy() for i in range(4 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = Text("GPU" , font_size=24 ) _snake_case : Optional[int] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = Text("Model" , font_size=24 ) _snake_case : Dict = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case : Union[str, Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : Union[str, Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Union[str, Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Optional[Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _snake_case : List[Any] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : int = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Dict = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) _snake_case : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
670
1
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = DanceDiffusionPipeline _lowerCamelCase = UNCONDITIONAL_AUDIO_GENERATION_PARAMS _lowerCamelCase = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } _lowerCamelCase = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDModel( block_out_channels=(32, 32, 64) , extra_in_channels=16 , sample_size=512 , sample_rate=16_000 , in_channels=2 , out_channels=2 , flip_sin_to_cos=lowercase_ , use_timestep_embedding=lowercase_ , time_embedding_type="fourier" , mid_block_type="UNetMidBlock1D" , down_block_types=("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , up_block_types=("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , ) _snake_case : Any = IPNDMScheduler() _snake_case : int = { "unet": unet, "scheduler": scheduler, } return components def UpperCamelCase ( self , lowercase_ , lowercase_=0 ): if str(lowercase_ ).startswith("mps" ): _snake_case : int = torch.manual_seed(lowercase_ ) else: _snake_case : List[Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) _snake_case : Tuple = { "batch_size": 1, "generator": generator, "num_inference_steps": 4, } return inputs def UpperCamelCase ( self ): _snake_case : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : List[Any] = self.get_dummy_components() _snake_case : Any = DanceDiffusionPipeline(**lowercase_ ) _snake_case : List[str] = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) _snake_case : Dict = self.get_dummy_inputs(lowercase_ ) _snake_case : List[str] = pipe(**lowercase_ ) _snake_case : Optional[int] = output.audios _snake_case : Any = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) _snake_case : List[Any] = np.array([-0.7_265, 1.0_000, -0.8_388, 0.1_175, 0.9_498, -1.0_000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def UpperCamelCase ( self ): return super().test_save_load_local() @skip_mps def UpperCamelCase ( self ): return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def UpperCamelCase ( self ): return super().test_save_load_optional_components() @skip_mps def UpperCamelCase ( self ): return super().test_attention_slicing_forward_pass() def UpperCamelCase ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self ): _snake_case : str = torch_device _snake_case : Dict = DanceDiffusionPipeline.from_pretrained("harmonai/maestro-150k" ) _snake_case : int = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) _snake_case : Optional[int] = torch.manual_seed(0 ) _snake_case : List[str] = pipe(generator=lowercase_ , num_inference_steps=100 , audio_length_in_s=4.096 ) _snake_case : Dict = output.audios _snake_case : Dict = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _snake_case : Optional[int] = np.array([-0.0_192, -0.0_231, -0.0_318, -0.0_059, 0.0_002, -0.0_020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = torch_device _snake_case : List[str] = DanceDiffusionPipeline.from_pretrained("harmonai/maestro-150k" , torch_dtype=torch.floataa ) _snake_case : int = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) _snake_case : Dict = torch.manual_seed(0 ) _snake_case : Any = pipe(generator=lowercase_ , num_inference_steps=100 , audio_length_in_s=4.096 ) _snake_case : Optional[Any] = output.audios _snake_case : List[Any] = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _snake_case : str = np.array([-0.0_367, -0.0_488, -0.0_771, -0.0_525, -0.0_444, -0.0_341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
670
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
1
import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) __SCREAMING_SNAKE_CASE : str = logging.getLogger() def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : List[Any] = {} _snake_case : List[str] = os.path.join(__lowercase , "all_results.json" ) if os.path.exists(__lowercase ): with open(__lowercase , "r" ) as f: _snake_case : int = json.load(__lowercase ) else: raise ValueError(F"""can't find {path}""" ) return results __SCREAMING_SNAKE_CASE : str = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): import xla_spawn _snake_case : Optional[Any] = self.get_auto_remove_tmp_dir() _snake_case : List[Any] = f""" ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(lowercase_ , "argv" , lowercase_ ): _snake_case : Optional[Any] = time() xla_spawn.main() _snake_case : Dict = time() _snake_case : Optional[int] = get_results(lowercase_ ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def UpperCamelCase ( self ): import xla_spawn _snake_case : Union[str, Any] = "\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n ".split() with patch.object(lowercase_ , "argv" , lowercase_ ): xla_spawn.main()
670
from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowercase_ ( __snake_case ): _lowerCamelCase = 'roc_bert' def __init__( self , lowercase_=30_522 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3_072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1e-12 , lowercase_=True , lowercase_=0 , lowercase_="absolute" , lowercase_=None , lowercase_=True , lowercase_=True , lowercase_=768 , lowercase_=910 , lowercase_=512 , lowercase_=24_858 , lowercase_=True , **lowercase_ , ): _snake_case : int = vocab_size _snake_case : Union[str, Any] = max_position_embeddings _snake_case : Union[str, Any] = hidden_size _snake_case : Dict = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : List[Any] = hidden_act _snake_case : Optional[int] = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : Union[str, Any] = initializer_range _snake_case : List[Any] = type_vocab_size _snake_case : int = layer_norm_eps _snake_case : Optional[Any] = use_cache _snake_case : List[Any] = enable_pronunciation _snake_case : Dict = enable_shape _snake_case : Dict = pronunciation_embed_dim _snake_case : Tuple = pronunciation_vocab_size _snake_case : Tuple = shape_embed_dim _snake_case : List[str] = shape_vocab_size _snake_case : Dict = concat_input _snake_case : int = position_embedding_type _snake_case : int = classifier_dropout super().__init__(pad_token_id=lowercase_ , **lowercase_ )
670
1
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def snake_case (__lowercase , __lowercase , __lowercase ) -> Tuple: '''simple docstring''' _snake_case : str = TaConfig.from_json_file(__lowercase ) print(F"""Building PyTorch model from configuration: {config}""" ) _snake_case : int = TaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tf_weights_in_ta(__lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(__lowercase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
670
from cva import destroyAllWindows, imread, imshow, waitKey def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case ,_snake_case : int = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(__lowercase ): for j in range(__lowercase ): _snake_case : Optional[Any] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image __SCREAMING_SNAKE_CASE : Optional[Any] = imread('image_data/lena.jpg', 1) # convert to its negative __SCREAMING_SNAKE_CASE : Tuple = convert_to_negative(img) # show result image imshow('negative of original image', img) waitKey(0) destroyAllWindows()
670
1
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def snake_case (__lowercase ) -> Dict[str, torch.Tensor]: '''simple docstring''' _snake_case : int = [] _snake_case : Optional[int] = [] _snake_case : str = [] for rt in rc.restypes: _snake_case : Dict = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) _snake_case : List[Any] = {name: i for i, name in enumerate(__lowercase )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) _snake_case : List[Any] = torch.tensor( __lowercase , dtype=torch.intaa , device=protein["aatype"].device , ) _snake_case : List[Any] = torch.tensor( __lowercase , dtype=torch.intaa , device=protein["aatype"].device , ) _snake_case : Union[str, Any] = torch.tensor( __lowercase , dtype=torch.floataa , device=protein["aatype"].device , ) _snake_case : str = protein["aatype"].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein _snake_case : Any = restype_atomaa_to_atomaa[protein_aatype] _snake_case : Dict = restype_atomaa_mask[protein_aatype] _snake_case : Dict = residx_atomaa_mask _snake_case : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back _snake_case : int = restype_atomaa_to_atomaa[protein_aatype] _snake_case : Any = residx_atomaa_to_atomaa.long() # create the corresponding mask _snake_case : Tuple = torch.zeros([21, 37] , dtype=torch.floataa , device=protein["aatype"].device ) for restype, restype_letter in enumerate(rc.restypes ): _snake_case : int = rc.restype_atoa[restype_letter] _snake_case : Union[str, Any] = rc.residue_atoms[restype_name] for atom_name in atom_names: _snake_case : Dict = rc.atom_order[atom_name] _snake_case : Any = 1 _snake_case : List[str] = restype_atomaa_mask[protein_aatype] _snake_case : List[Any] = residx_atomaa_mask return protein def snake_case (__lowercase ) -> Dict[str, np.ndarray]: '''simple docstring''' _snake_case : Optional[int] = tree_map(lambda __lowercase : torch.tensor(__lowercase , device=batch["aatype"].device ) , __lowercase , np.ndarray ) _snake_case : Any = tensor_tree_map(lambda __lowercase : np.array(__lowercase ) , make_atomaa_masks(__lowercase ) ) return out
670
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __SCREAMING_SNAKE_CASE : List[str] = Mapping[str, np.ndarray] __SCREAMING_SNAKE_CASE : List[Any] = Mapping[str, Any] # Is a nested dict. __SCREAMING_SNAKE_CASE : List[Any] = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class lowercase_ : _lowerCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. _lowerCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. _lowerCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. _lowerCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. _lowerCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions _lowerCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files _lowerCamelCase = None # Templates used to generate this protein (prediction-only) _lowerCamelCase = None # Chain corresponding to each parent _lowerCamelCase = None def snake_case (__lowercase ) -> Protein: '''simple docstring''' _snake_case : str = r"(\[[A-Z]+\]\n)" _snake_case : List[str] = [tag.strip() for tag in re.split(__lowercase , __lowercase ) if len(__lowercase ) > 0] _snake_case : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("\n" ) for l in tags[1::2]] ) _snake_case : List[str] = ["N", "CA", "C"] _snake_case : Any = None _snake_case : Union[str, Any] = None _snake_case : Optional[int] = None for g in groups: if "[PRIMARY]" == g[0]: _snake_case : Tuple = g[1][0].strip() for i in range(len(__lowercase ) ): if seq[i] not in residue_constants.restypes: _snake_case : Tuple = "X" # FIXME: strings are immutable _snake_case : int = np.array( [residue_constants.restype_order.get(__lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: _snake_case : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(__lowercase , g[1][axis].split() ) ) ) _snake_case : Dict = np.array(__lowercase ) _snake_case : Dict = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: _snake_case : int = np.array(list(map({"-": 0, "+": 1}.get , g[1][0].strip() ) ) ) _snake_case : Any = np.zeros( ( len(__lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__lowercase ): _snake_case : Dict = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__lowercase , atom_mask=__lowercase , aatype=__lowercase , residue_index=np.arange(len(__lowercase ) ) , b_factors=__lowercase , ) def snake_case (__lowercase , __lowercase = 0 ) -> List[str]: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[Any] = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) _snake_case : str = prot.parents _snake_case : str = prot.parents_chain_index if parents is not None and parents_chain_index is not None: _snake_case : int = [p for i, p in zip(__lowercase , __lowercase ) if i == chain_id] if parents is None or len(__lowercase ) == 0: _snake_case : Optional[int] = ["N/A"] pdb_headers.append(F"""PARENT {' '.join(__lowercase )}""" ) return pdb_headers def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : List[str] = [] _snake_case : Optional[int] = pdb_str.split("\n" ) _snake_case : List[str] = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) _snake_case : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: _snake_case : str = [] if prot.parents_chain_index is not None: _snake_case : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(__lowercase ) , [] ) parent_dict[str(__lowercase )].append(__lowercase ) _snake_case : Any = max([int(__lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): _snake_case : Tuple = parent_dict.get(str(__lowercase ) , ["N/A"] ) parents_per_chain.append(__lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: _snake_case : List[str] = [["N/A"]] def make_parent_line(__lowercase ) -> str: return F"""PARENT {' '.join(__lowercase )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) _snake_case : int = 0 for i, l in enumerate(__lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__lowercase ): _snake_case : Tuple = parents_per_chain[chain_counter] else: _snake_case : str = ["N/A"] out_pdb_lines.append(make_parent_line(__lowercase ) ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Optional[Any] = residue_constants.restypes + ["X"] def res_atoa(__lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , "UNK" ) _snake_case : Optional[int] = residue_constants.atom_types _snake_case : List[str] = [] _snake_case : Tuple = prot.atom_mask _snake_case : List[str] = prot.aatype _snake_case : int = prot.atom_positions _snake_case : int = prot.residue_index.astype(np.intaa ) _snake_case : List[Any] = prot.b_factors _snake_case : str = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("Invalid aatypes." ) _snake_case : Union[str, Any] = get_pdb_headers(__lowercase ) if len(__lowercase ) > 0: pdb_lines.extend(__lowercase ) _snake_case : Optional[Any] = aatype.shape[0] _snake_case : str = 1 _snake_case : Tuple = 0 _snake_case : int = string.ascii_uppercase _snake_case : Optional[Any] = None # Add all atom sites. for i in range(__lowercase ): _snake_case : Dict = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue _snake_case : List[Any] = "ATOM" _snake_case : Union[str, Any] = atom_name if len(__lowercase ) == 4 else F""" {atom_name}""" _snake_case : str = "" _snake_case : str = "" _snake_case : Any = 1.00 _snake_case : str = atom_name[0] # Protein supports only C, N, O, S, this works. _snake_case : Dict = "" _snake_case : Any = "A" if chain_index is not None: _snake_case : List[Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! _snake_case : Optional[int] = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(__lowercase ) atom_index += 1 _snake_case : Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: _snake_case : Optional[int] = True _snake_case : Union[str, Any] = chain_index[i + 1] if should_terminate: # Close the chain. _snake_case : List[str] = "TER" _snake_case : str = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(__lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__lowercase , __lowercase ) ) pdb_lines.append("END" ) pdb_lines.append("" ) return "\n".join(__lowercase ) def snake_case (__lowercase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , ) -> Protein: '''simple docstring''' return Protein( aatype=features["aatype"] , atom_positions=result["final_atom_positions"] , atom_mask=result["final_atom_mask"] , residue_index=features["residue_index"] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["final_atom_mask"] ) , chain_index=__lowercase , remark=__lowercase , parents=__lowercase , parents_chain_index=__lowercase , )
670
1
import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) __SCREAMING_SNAKE_CASE : int = logging.getLogger() def snake_case (__lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' _snake_case : List[Any] = "\n".join(__lowercase ) Path(__lowercase ).open("w" ).writelines(__lowercase ) __SCREAMING_SNAKE_CASE : int = 'patrickvonplaten/t5-tiny-random' __SCREAMING_SNAKE_CASE : List[str] = 'sshleifer/bart-tiny-random' __SCREAMING_SNAKE_CASE : Optional[Any] = 'sshleifer/tiny-mbart' __SCREAMING_SNAKE_CASE : Any = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowercase_ ( __snake_case ): def UpperCamelCase ( self , lowercase_ ): _snake_case : Tuple = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" _snake_case : int = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() _snake_case : Optional[Any] = [" New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County."] _dump_articles(lowercase_ , lowercase_ ) _snake_case : Dict = str(Path(self.get_auto_remove_tmp_dir() ) / "scores.json" ) _snake_case : int = "translation_en_to_de" if model == T5_TINY else "summarization" _snake_case : int = f""" run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 """.split() with patch.object(lowercase_ , "argv" , lowercase_ ): run_generate() assert Path(lowercase_ ).exists() # os.remove(Path(output_file_name)) def UpperCamelCase ( self ): self.run_eval_tester(lowercase_ ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def UpperCamelCase ( self , lowercase_ ): self.run_eval_tester(lowercase_ ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def UpperCamelCase ( self , lowercase_ ): _snake_case : Any = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" _snake_case : Dict = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() _snake_case : int = { "en": ["Machine learning is great, isn't it?", "I like to eat bananas", "Tomorrow is another great day!"], "de": [ "Maschinelles Lernen ist großartig, oder?", "Ich esse gerne Bananen", "Morgen ist wieder ein toller Tag!", ], } _snake_case : List[str] = Path(self.get_auto_remove_tmp_dir() ) _snake_case : Union[str, Any] = str(tmp_dir / "scores.json" ) _snake_case : Any = str(tmp_dir / "val.target" ) _dump_articles(lowercase_ , text["en"] ) _dump_articles(lowercase_ , text["de"] ) _snake_case : str = "translation_en_to_de" if model == T5_TINY else "summarization" _snake_case : Union[str, Any] = f""" run_eval_search.py {model} {str(lowercase_ )} {str(lowercase_ )} --score_path {score_path} --reference_path {reference_path} --task {task} """.split() testargs.extend(["--search", "num_beams=1:2 length_penalty=0.9:1.0"] ) with patch.object(lowercase_ , "argv" , lowercase_ ): with CaptureStdout() as cs: run_search() _snake_case : Optional[int] = [" num_beams | length_penalty", model, "Best score args"] _snake_case : Optional[Any] = ["Info"] if "translation" in task: expected_strings.append("bleu" ) else: expected_strings.extend(lowercase_ ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(lowercase_ ).exists() os.remove(Path(lowercase_ ) )
670
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor'] _lowerCamelCase = 'SamImageProcessor' def __init__( self , lowercase_ ): super().__init__(lowercase_ ) _snake_case : Optional[Any] = self.image_processor _snake_case : Tuple = -10 _snake_case : str = self.image_processor.size["longest_edge"] def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_ = None , **lowercase_ , ): _snake_case : List[Any] = self.image_processor( lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # pop arguments that are not used in the foward but used nevertheless _snake_case : Any = encoding_image_processor["original_sizes"] if hasattr(lowercase_ , "numpy" ): # Checks if Torch or TF tensor _snake_case : int = original_sizes.numpy() _snake_case ,_snake_case ,_snake_case : Union[str, Any] = self._check_and_preprocess_points( input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , ) _snake_case : Dict = self._normalize_and_convert( lowercase_ , lowercase_ , input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , return_tensors=lowercase_ , ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_="pt" , ): if input_points is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : int = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] ) for point in input_points ] else: _snake_case : Dict = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ ) for point, original_size in zip(lowercase_ , lowercase_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: _snake_case ,_snake_case : int = self._pad_points_and_labels(lowercase_ , lowercase_ ) _snake_case : Any = np.array(lowercase_ ) if input_labels is not None: _snake_case : Optional[Any] = np.array(lowercase_ ) if input_boxes is not None: if len(lowercase_ ) != len(lowercase_ ): _snake_case : Optional[Any] = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] , is_bounding_box=lowercase_ ) for box in input_boxes ] else: _snake_case : List[str] = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ , is_bounding_box=lowercase_ ) for box, original_size in zip(lowercase_ , lowercase_ ) ] _snake_case : Tuple = np.array(lowercase_ ) if input_boxes is not None: if return_tensors == "pt": _snake_case : List[str] = torch.from_numpy(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # boxes batch size of 1 by default _snake_case : Optional[int] = tf.expand_dims(lowercase_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": _snake_case : Tuple = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : int = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": _snake_case : List[str] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : Tuple = tf.expand_dims(lowercase_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": _snake_case : Dict = torch.from_numpy(lowercase_ ) # point batch size of 1 by default _snake_case : str = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": _snake_case : Optional[Any] = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default _snake_case : List[Any] = tf.expand_dims(lowercase_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : List[Any] = max([point.shape[0] for point in input_points] ) _snake_case : List[str] = [] for i, point in enumerate(lowercase_ ): if point.shape[0] != expected_nb_points: _snake_case : Optional[Any] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) _snake_case : Union[str, Any] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowercase_ ) _snake_case : Optional[Any] = processed_input_points return input_points, input_labels def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=False ): _snake_case ,_snake_case : Optional[int] = original_size _snake_case ,_snake_case : List[str] = self.image_processor._get_preprocess_shape(lowercase_ , longest_edge=lowercase_ ) _snake_case : Optional[Any] = deepcopy(lowercase_ ).astype(lowercase_ ) if is_bounding_box: _snake_case : str = coords.reshape(-1 , 2 , 2 ) _snake_case : Optional[Any] = coords[..., 0] * (new_w / old_w) _snake_case : Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: _snake_case : Optional[Any] = coords.reshape(-1 , 4 ) return coords def UpperCamelCase ( self , lowercase_=None , lowercase_=None , lowercase_=None , ): if input_points is not None: if hasattr(lowercase_ , "numpy" ): # Checks for TF or Torch tensor _snake_case : Union[str, Any] = input_points.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_points[0] , lowercase_ ): raise ValueError("Input points must be a list of list of floating points." ) _snake_case : Any = [np.array(lowercase_ ) for input_point in input_points] else: _snake_case : Optional[int] = None if input_labels is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : Tuple = input_labels.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_labels[0] , lowercase_ ): raise ValueError("Input labels must be a list of list integers." ) _snake_case : Tuple = [np.array(lowercase_ ) for label in input_labels] else: _snake_case : Optional[Any] = None if input_boxes is not None: if hasattr(lowercase_ , "numpy" ): _snake_case : List[str] = input_boxes.numpy().tolist() if ( not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_boxes[0] , lowercase_ ) or not isinstance(input_boxes[0][0] , lowercase_ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) _snake_case : List[Any] = [np.array(lowercase_ ).astype(np.floataa ) for box in input_boxes] else: _snake_case : Optional[int] = None return input_points, input_labels, input_boxes @property def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(lowercase_ ) ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.image_processor.post_process_masks(*lowercase_ , **lowercase_ )
670
1
import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) def snake_case (__lowercase , __lowercase , __lowercase ) -> Dict: '''simple docstring''' _snake_case : Optional[Any] = UniSpeechSatForSequenceClassification.from_pretrained(__lowercase , config=__lowercase ) _snake_case : Union[str, Any] = downstream_dict["projector.weight"] _snake_case : Dict = downstream_dict["projector.bias"] _snake_case : Dict = downstream_dict["model.post_net.linear.weight"] _snake_case : str = downstream_dict["model.post_net.linear.bias"] return model def snake_case (__lowercase , __lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' _snake_case : Union[str, Any] = UniSpeechSatForAudioFrameClassification.from_pretrained(__lowercase , config=__lowercase ) _snake_case : Any = downstream_dict["model.linear.weight"] _snake_case : Optional[Any] = downstream_dict["model.linear.bias"] return model def snake_case (__lowercase , __lowercase , __lowercase ) -> Tuple: '''simple docstring''' _snake_case : List[Any] = UniSpeechSatForXVector.from_pretrained(__lowercase , config=__lowercase ) _snake_case : Tuple = downstream_dict["connector.weight"] _snake_case : int = downstream_dict["connector.bias"] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): _snake_case : Optional[Any] = downstream_dict[ F"""model.framelevel_feature_extractor.module.{i}.kernel.weight""" ] _snake_case : Union[str, Any] = downstream_dict[F"""model.framelevel_feature_extractor.module.{i}.kernel.bias"""] _snake_case : Union[str, Any] = downstream_dict["model.utterancelevel_feature_extractor.linear1.weight"] _snake_case : Any = downstream_dict["model.utterancelevel_feature_extractor.linear1.bias"] _snake_case : List[Any] = downstream_dict["model.utterancelevel_feature_extractor.linear2.weight"] _snake_case : Optional[Any] = downstream_dict["model.utterancelevel_feature_extractor.linear2.bias"] _snake_case : str = downstream_dict["objective.W"] return model @torch.no_grad() def snake_case (__lowercase , __lowercase , __lowercase , __lowercase ) -> Any: '''simple docstring''' _snake_case : List[str] = torch.load(__lowercase , map_location="cpu" ) _snake_case : List[str] = checkpoint["Downstream"] _snake_case : Optional[int] = UniSpeechSatConfig.from_pretrained(__lowercase ) _snake_case : int = WavaVecaFeatureExtractor.from_pretrained( __lowercase , return_attention_mask=__lowercase , do_normalize=__lowercase ) _snake_case : Tuple = hf_config.architectures[0] if arch.endswith("ForSequenceClassification" ): _snake_case : str = convert_classification(__lowercase , __lowercase , __lowercase ) elif arch.endswith("ForAudioFrameClassification" ): _snake_case : Any = convert_diarization(__lowercase , __lowercase , __lowercase ) elif arch.endswith("ForXVector" ): _snake_case : List[str] = convert_xvector(__lowercase , __lowercase , __lowercase ) else: raise NotImplementedError(F"""S3PRL weights conversion is not supported for {arch}""" ) if hf_config.use_weighted_layer_sum: _snake_case : Dict = checkpoint["Featurizer"]["weights"] hf_feature_extractor.save_pretrained(__lowercase ) hf_model.save_pretrained(__lowercase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') __SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
670
def snake_case (__lowercase ) -> int: '''simple docstring''' if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _snake_case : Union[str, Any] = grid[0] for row_n in range(1 , len(__lowercase ) ): _snake_case : Union[str, Any] = grid[row_n] _snake_case : List[Any] = fill_row(__lowercase , __lowercase ) _snake_case : List[Any] = grid[row_n] return grid[-1][-1] def snake_case (__lowercase , __lowercase ) -> list: '''simple docstring''' current_row[0] += row_above[0] for cell_n in range(1 , len(__lowercase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
670
1
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def snake_case (__lowercase ) -> Tuple: '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : Optional[Any] = np.max(_outputs , axis=-1 , keepdims=__lowercase ) _snake_case : int = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=__lowercase ) class lowercase_ ( __snake_case ): _lowerCamelCase = 'sigmoid' _lowerCamelCase = 'softmax' _lowerCamelCase = 'none' @add_end_docstrings( __snake_case , R'\n return_all_scores (`bool`, *optional*, defaults to `False`):\n Whether to return all prediction scores or just the one of the predicted class.\n function_to_apply (`str`, *optional*, defaults to `"default"`):\n The function to apply to the model outputs in order to retrieve the scores. Accepts four different values:\n\n - `"default"`: if the model has a single label, will apply the sigmoid function on the output. If the model\n has several labels, will apply the softmax function on the output.\n - `"sigmoid"`: Applies the sigmoid function on the output.\n - `"softmax"`: Applies the softmax function on the output.\n - `"none"`: Does not apply any function on the output.\n ' , ) class lowercase_ ( __snake_case ): _lowerCamelCase = False _lowerCamelCase = ClassificationFunction.NONE def __init__( self , **lowercase_ ): super().__init__(**lowercase_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == "tf" else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def UpperCamelCase ( self , lowercase_=None , lowercase_=None , lowercase_="" , **lowercase_ ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" _snake_case : List[str] = tokenizer_kwargs _snake_case : List[str] = {} if hasattr(self.model.config , "return_all_scores" ) and return_all_scores is None: _snake_case : Union[str, Any] = self.model.config.return_all_scores if isinstance(lowercase_ , lowercase_ ) or top_k is None: _snake_case : List[str] = top_k _snake_case : Union[str, Any] = False elif return_all_scores is not None: warnings.warn( "`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of" " `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`." , lowercase_ , ) if return_all_scores: _snake_case : int = None else: _snake_case : List[Any] = 1 if isinstance(lowercase_ , lowercase_ ): _snake_case : str = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: _snake_case : Tuple = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *lowercase_ , **lowercase_ ): _snake_case : List[str] = super().__call__(*lowercase_ , **lowercase_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. _snake_case : Any = "top_k" not in kwargs if isinstance(args[0] , lowercase_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def UpperCamelCase ( self , lowercase_ , **lowercase_ ): _snake_case : Any = self.framework if isinstance(lowercase_ , lowercase_ ): return self.tokenizer(**lowercase_ , return_tensors=lowercase_ , **lowercase_ ) elif isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) == 1 and isinstance(inputs[0] , lowercase_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=lowercase_ , **lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( "The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a" " dictionary `{\"text\": \"My text\", \"text_pair\": \"My pair\"}` in order to send a text pair." ) return self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ ): return self.model(**lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_=None , lowercase_=1 , lowercase_=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: _snake_case : Optional[Any] = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: _snake_case : Union[str, Any] = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , "function_to_apply" ) and function_to_apply is None: _snake_case : Dict = self.model.config.function_to_apply else: _snake_case : Optional[Any] = ClassificationFunction.NONE _snake_case : Optional[Any] = model_outputs["logits"][0] _snake_case : int = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: _snake_case : Dict = sigmoid(lowercase_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: _snake_case : Tuple = softmax(lowercase_ ) elif function_to_apply == ClassificationFunction.NONE: _snake_case : List[Any] = outputs else: raise ValueError(f"""Unrecognized `function_to_apply` argument: {function_to_apply}""" ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} _snake_case : Optional[Any] = [ {"label": self.model.config.idalabel[i], "score": score.item()} for i, score in enumerate(lowercase_ ) ] if not _legacy: dict_scores.sort(key=lambda lowercase_ : x["score"] , reverse=lowercase_ ) if top_k is not None: _snake_case : Tuple = dict_scores[:top_k] return dict_scores
670
import random def snake_case (__lowercase , __lowercase ) -> tuple: '''simple docstring''' _snake_case ,_snake_case ,_snake_case : List[Any] = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def snake_case (__lowercase , __lowercase ) -> List[Any]: '''simple docstring''' if index >= len(__lowercase ) or index < 0: return None _snake_case : Any = items[random.randint(0 , len(__lowercase ) - 1 )] _snake_case : Tuple = 0 _snake_case ,_snake_case ,_snake_case : Tuple = _partition(__lowercase , __lowercase ) _snake_case : Tuple = len(__lowercase ) _snake_case : List[str] = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
670
1
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( 'stable diffusion controlnet', '0.22.0', 'Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.', standard_warn=False, stacklevel=3, )
670
from math import pow, sqrt def snake_case (*__lowercase ) -> bool: '''simple docstring''' _snake_case : str = len(__lowercase ) > 0 and all(value > 0.0 for value in values ) return result def snake_case (__lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase ) else ValueError("Input Error: Molar mass values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) ) def snake_case (__lowercase , __lowercase , __lowercase ) -> float | ValueError: '''simple docstring''' return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(__lowercase , __lowercase , __lowercase ) else ValueError( "Input Error: Molar mass and effusion rate values must greater than 0." ) )
670
1
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = False , lowercase_ = None , lowercase_ = None , **lowercase_ , ): super().__init__( lowercase_ , split=lowercase_ , features=lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ , streaming=lowercase_ , num_proc=lowercase_ , **lowercase_ , ) _snake_case : Tuple = field _snake_case : Union[str, Any] = path_or_paths if isinstance(lowercase_ , lowercase_ ) else {self.split: path_or_paths} _snake_case : Optional[int] = Json( cache_dir=lowercase_ , data_files=lowercase_ , features=lowercase_ , field=lowercase_ , **lowercase_ , ) def UpperCamelCase ( self ): # Build iterable dataset if self.streaming: _snake_case : Dict = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _snake_case : Optional[Any] = None _snake_case : int = None _snake_case : Tuple = None _snake_case : Optional[Any] = None self.builder.download_and_prepare( download_config=lowercase_ , download_mode=lowercase_ , verification_mode=lowercase_ , base_path=lowercase_ , num_proc=self.num_proc , ) _snake_case : int = self.builder.as_dataset( split=self.split , verification_mode=lowercase_ , in_memory=self.keep_in_memory ) return dataset class lowercase_ : def __init__( self , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , **lowercase_ , ): if num_proc is not None and num_proc <= 0: raise ValueError(f"""num_proc {num_proc} must be an integer > 0.""" ) _snake_case : Dict = dataset _snake_case : List[Any] = path_or_buf _snake_case : List[str] = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE _snake_case : int = num_proc _snake_case : Optional[int] = "utf-8" _snake_case : List[Any] = to_json_kwargs def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.to_json_kwargs.pop("path_or_buf" , lowercase_ ) _snake_case : Dict = self.to_json_kwargs.pop("orient" , "records" ) _snake_case : Any = self.to_json_kwargs.pop("lines" , True if orient == "records" else False ) _snake_case : List[str] = self.to_json_kwargs.pop("index" , False if orient in ["split", "table"] else True ) _snake_case : Union[str, Any] = self.to_json_kwargs.pop("compression" , lowercase_ ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(f"""`datasets` currently does not support {compression} compression""" ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , "wb" , compression=lowercase_ ) as buffer: _snake_case : Optional[int] = self._write(file_obj=lowercase_ , orient=lowercase_ , lines=lowercase_ , index=lowercase_ , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( f"""The compression parameter is not supported when writing to a buffer, but compression={compression}""" " was passed. Please provide a local path instead." ) _snake_case : Tuple = self._write( file_obj=self.path_or_buf , orient=lowercase_ , lines=lowercase_ , index=lowercase_ , **self.to_json_kwargs ) return written def UpperCamelCase ( self , lowercase_ ): _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case : Any = args _snake_case : Any = query_table( table=self.dataset.data , key=slice(lowercase_ , offset + self.batch_size ) , indices=self.dataset._indices , ) _snake_case : Dict = batch.to_pandas().to_json( path_or_buf=lowercase_ , orient=lowercase_ , lines=lowercase_ , index=lowercase_ , **lowercase_ ) if not json_str.endswith("\n" ): json_str += "\n" return json_str.encode(self.encoding ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , **lowercase_ , ): _snake_case : Union[str, Any] = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): _snake_case : List[str] = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(lowercase_ ) else: _snake_case ,_snake_case : Union[str, Any] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , lowercase_ , lowercase_ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating json from Arrow format" , ): written += file_obj.write(lowercase_ ) return written
670
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , *lowercase_ , **lowercase_ ): warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
670
1
import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __SCREAMING_SNAKE_CASE : Optional[int] = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.14.0', 'To fix: pip install -r examples/pytorch/audio-classification/requirements.txt') def snake_case (__lowercase , __lowercase , __lowercase = 16_000 ) -> Any: '''simple docstring''' _snake_case : str = int(round(sample_rate * max_length ) ) if len(__lowercase ) <= sample_length: return wav _snake_case : Union[str, Any] = randint(0 , len(__lowercase ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class lowercase_ : _lowerCamelCase = field(default=__snake_case , metadata={'help': 'Name of a dataset from the datasets package'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'A file containing the training audio paths and labels.'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'A file containing the validation audio paths and labels.'} ) _lowerCamelCase = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) _lowerCamelCase = field( default='validation' , metadata={ 'help': ( 'The name of the training data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) _lowerCamelCase = field( default='audio' , metadata={'help': 'The name of the dataset column containing the audio data. Defaults to \'audio\''} , ) _lowerCamelCase = field( default='label' , metadata={'help': 'The name of the dataset column containing the labels. Defaults to \'label\''} ) _lowerCamelCase = field( default=__snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) _lowerCamelCase = field( default=__snake_case , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) _lowerCamelCase = field( default=20 , metadata={'help': 'Audio clips will be randomly cut to this length during training if the value is set.'} , ) @dataclass class lowercase_ : _lowerCamelCase = field( default='facebook/wav2vec2-base' , metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} , ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Where do you want to store the pretrained models downloaded from the Hub'} ) _lowerCamelCase = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Name or path of preprocessor config.'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Whether to freeze the feature encoder layers of the model.'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Whether to generate an attention mask in the feature extractor.'} ) _lowerCamelCase = field( default=__snake_case , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'} , ) def UpperCamelCase ( self ): if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( "The argument `--freeze_feature_extractor` is deprecated and " "will be removed in a future version. Use `--freeze_feature_encoder`" "instead. Setting `freeze_feature_encoder==True`." , lowercase_ , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( "The argument `--freeze_feature_extractor` is deprecated and " "should not be used in combination with `--freeze_feature_encoder`." "Only make use of `--freeze_feature_encoder`." ) def snake_case () -> int: '''simple docstring''' _snake_case : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _snake_case ,_snake_case ,_snake_case : Optional[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _snake_case ,_snake_case ,_snake_case : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_audio_classification" , __lowercase , __lowercase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _snake_case : int = training_args.get_process_log_level() logger.setLevel(__lowercase ) transformers.utils.logging.set_verbosity(__lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} """ + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _snake_case : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _snake_case : List[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to train from scratch." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Initialize our dataset and prepare it for the audio classification task. _snake_case : int = DatasetDict() _snake_case : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _snake_case : Optional[int] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--audio_column_name {data_args.audio_column_name} not found in dataset '{data_args.dataset_name}'. """ "Make sure to set `--audio_column_name` to the correct audio column - one of " F"""{', '.join(raw_datasets['train'].column_names )}.""" ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--label_column_name {data_args.label_column_name} not found in dataset '{data_args.dataset_name}'. """ "Make sure to set `--label_column_name` to the correct text column - one of " F"""{', '.join(raw_datasets['train'].column_names )}.""" ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _snake_case : Dict = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _snake_case : Optional[int] = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _snake_case : Optional[int] = feature_extractor.model_input_names[0] def train_transforms(__lowercase ): _snake_case : Dict = [] for audio in batch[data_args.audio_column_name]: _snake_case : Any = random_subsample( audio["array"] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(__lowercase ) _snake_case : Dict = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _snake_case : List[Any] = {model_input_name: inputs.get(__lowercase )} _snake_case : Dict = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(__lowercase ): _snake_case : Tuple = [audio["array"] for audio in batch[data_args.audio_column_name]] _snake_case : Union[str, Any] = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _snake_case : int = {model_input_name: inputs.get(__lowercase )} _snake_case : Any = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _snake_case : Optional[int] = raw_datasets["train"].features[data_args.label_column_name].names _snake_case ,_snake_case : List[Any] = {}, {} for i, label in enumerate(__lowercase ): _snake_case : Dict = str(__lowercase ) _snake_case : List[str] = label # Load the accuracy metric from the datasets package _snake_case : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(__lowercase ): _snake_case : str = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=__lowercase , references=eval_pred.label_ids ) _snake_case : int = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowercase ) , labelaid=__lowercase , idalabel=__lowercase , finetuning_task="audio-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _snake_case : Union[str, Any] = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _snake_case : Optional[Any] = ( raw_datasets["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(__lowercase , output_all_columns=__lowercase ) if training_args.do_eval: if data_args.max_eval_samples is not None: _snake_case : List[Any] = ( raw_datasets["eval"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(__lowercase , output_all_columns=__lowercase ) # Initialize our trainer _snake_case : Tuple = Trainer( model=__lowercase , args=__lowercase , train_dataset=raw_datasets["train"] if training_args.do_train else None , eval_dataset=raw_datasets["eval"] if training_args.do_eval else None , compute_metrics=__lowercase , tokenizer=__lowercase , ) # Training if training_args.do_train: _snake_case : Optional[Any] = None if training_args.resume_from_checkpoint is not None: _snake_case : Dict = training_args.resume_from_checkpoint elif last_checkpoint is not None: _snake_case : Optional[int] = last_checkpoint _snake_case : str = trainer.train(resume_from_checkpoint=__lowercase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _snake_case : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowercase ) trainer.save_metrics("eval" , __lowercase ) # Write model card and (optionally) push to hub _snake_case : Optional[int] = { "finetuned_from": model_args.model_name_or_path, "tasks": "audio-classification", "dataset": data_args.dataset_name, "tags": ["audio-classification"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowercase ) else: trainer.create_model_card(**__lowercase ) if __name__ == "__main__": main()
670
from __future__ import annotations from typing import TypedDict class lowercase_ ( __snake_case ): _lowerCamelCase = 42 _lowerCamelCase = 42 def snake_case (__lowercase ) -> list[str]: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) return [s[i:] + s[:i] for i in range(len(__lowercase ) )] def snake_case (__lowercase ) -> BWTTransformDict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter s type must be str." ) if not s: raise ValueError("The parameter s must not be empty." ) _snake_case : List[str] = all_rotations(__lowercase ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation _snake_case : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(__lowercase ), } return response def snake_case (__lowercase , __lowercase ) -> str: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("The parameter bwt_string type must be str." ) if not bwt_string: raise ValueError("The parameter bwt_string must not be empty." ) try: _snake_case : Union[str, Any] = int(__lowercase ) except ValueError: raise TypeError( "The parameter idx_original_string type must be int or passive" " of cast to int." ) if idx_original_string < 0: raise ValueError("The parameter idx_original_string must not be lower than 0." ) if idx_original_string >= len(__lowercase ): raise ValueError( "The parameter idx_original_string must be lower than" " len(bwt_string)." ) _snake_case : Optional[Any] = [""] * len(__lowercase ) for _ in range(len(__lowercase ) ): for i in range(len(__lowercase ) ): _snake_case : Tuple = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = 'Provide a string that I will generate its BWT transform: ' __SCREAMING_SNAKE_CASE : Optional[Any] = input(entry_msg).strip() __SCREAMING_SNAKE_CASE : int = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result['bwt_string']}\'''' ) __SCREAMING_SNAKE_CASE : List[str] = reverse_bwt(result['bwt_string'], result['idx_original_string']) print( F'''Reversing Burrows Wheeler transform for entry \'{result['bwt_string']}\' ''' F'''we get original string \'{original_string}\'''' )
670
1
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = MvpTokenizer _lowerCamelCase = MvpTokenizerFast _lowerCamelCase = True _lowerCamelCase = filter_roberta_detectors def UpperCamelCase ( self ): super().setUp() _snake_case : int = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] _snake_case : str = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : List[Any] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] _snake_case : Any = {"unk_token": "<unk>"} _snake_case : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) def UpperCamelCase ( self , **lowercase_ ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , lowercase_ ): return "lower newer", "lower newer" @cached_property def UpperCamelCase ( self ): return MvpTokenizer.from_pretrained("RUCAIBox/mvp" ) @cached_property def UpperCamelCase ( self ): return MvpTokenizerFast.from_pretrained("RUCAIBox/mvp" ) @require_torch def UpperCamelCase ( self ): _snake_case : Optional[int] = ["A long paragraph for summarization.", "Another paragraph for summarization."] _snake_case : Any = [0, 250, 251, 17_818, 13, 39_186, 1_938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _snake_case : Optional[Any] = tokenizer(lowercase_ , max_length=len(lowercase_ ) , padding=lowercase_ , return_tensors="pt" ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) _snake_case : int = batch.input_ids.tolist()[0] self.assertListEqual(lowercase_ , lowercase_ ) # Test that special tokens are reset @require_torch def UpperCamelCase ( self ): _snake_case : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _snake_case : int = tokenizer(lowercase_ , padding=lowercase_ , return_tensors="pt" ) # check if input_ids are returned and no labels self.assertIn("input_ids" , lowercase_ ) self.assertIn("attention_mask" , lowercase_ ) self.assertNotIn("labels" , lowercase_ ) self.assertNotIn("decoder_attention_mask" , lowercase_ ) @require_torch def UpperCamelCase ( self ): _snake_case : Optional[int] = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _snake_case : Dict = tokenizer(text_target=lowercase_ , max_length=32 , padding="max_length" , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) @require_torch def UpperCamelCase ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _snake_case : str = tokenizer( ["I am a small frog" * 1_024, "I am a small frog"] , padding=lowercase_ , truncation=lowercase_ , return_tensors="pt" ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual(batch.input_ids.shape , (2, 1_024) ) @require_torch def UpperCamelCase ( self ): _snake_case : Optional[Any] = ["A long paragraph for summarization."] _snake_case : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _snake_case : List[str] = tokenizer(lowercase_ , text_target=lowercase_ , return_tensors="pt" ) _snake_case : List[Any] = inputs["input_ids"] _snake_case : Optional[int] = inputs["labels"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Any = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) _snake_case : List[Any] = self.tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) _snake_case : List[Any] = "A, <mask> AllenNLP sentence." _snake_case : str = tokenizer_r.encode_plus(lowercase_ , add_special_tokens=lowercase_ , return_token_type_ids=lowercase_ ) _snake_case : Tuple = tokenizer_p.encode_plus(lowercase_ , add_special_tokens=lowercase_ , return_token_type_ids=lowercase_ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) _snake_case : Tuple = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) _snake_case : str = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( lowercase_ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowercase_ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
670
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( 'stable diffusion controlnet', '0.22.0', 'Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.', standard_warn=False, stacklevel=3, )
670
1
import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel __SCREAMING_SNAKE_CASE : List[str] = '0.12' # assumed parallelism: 8 @require_flax @is_staging_test class lowercase_ ( unittest.TestCase ): @classmethod def UpperCamelCase ( cls ): _snake_case : Optional[Any] = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def UpperCamelCase ( cls ): try: delete_repo(token=cls._token , repo_id="test-model-flax" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-model-flax-org" ) except HTTPError: pass def UpperCamelCase ( self ): _snake_case : int = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) _snake_case : int = FlaxBertModel(lowercase_ ) model.push_to_hub("test-model-flax" , use_auth_token=self._token ) _snake_case : Optional[int] = FlaxBertModel.from_pretrained(f"""{USER}/test-model-flax""" ) _snake_case : List[Any] = flatten_dict(unfreeze(model.params ) ) _snake_case : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _snake_case : Tuple = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowercase_ , 1e-3 , msg=f"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id="test-model-flax" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase_ , repo_id="test-model-flax" , push_to_hub=lowercase_ , use_auth_token=self._token ) _snake_case : List[str] = FlaxBertModel.from_pretrained(f"""{USER}/test-model-flax""" ) _snake_case : Dict = flatten_dict(unfreeze(model.params ) ) _snake_case : Optional[int] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _snake_case : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowercase_ , 1e-3 , msg=f"""{key} not identical""" ) def UpperCamelCase ( self ): _snake_case : Tuple = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) _snake_case : Optional[int] = FlaxBertModel(lowercase_ ) model.push_to_hub("valid_org/test-model-flax-org" , use_auth_token=self._token ) _snake_case : List[Any] = FlaxBertModel.from_pretrained("valid_org/test-model-flax-org" ) _snake_case : Optional[Any] = flatten_dict(unfreeze(model.params ) ) _snake_case : Dict = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _snake_case : str = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowercase_ , 1e-3 , msg=f"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-model-flax-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( lowercase_ , repo_id="valid_org/test-model-flax-org" , push_to_hub=lowercase_ , use_auth_token=self._token ) _snake_case : Optional[int] = FlaxBertModel.from_pretrained("valid_org/test-model-flax-org" ) _snake_case : str = flatten_dict(unfreeze(model.params ) ) _snake_case : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _snake_case : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(lowercase_ , 1e-3 , msg=f"""{key} not identical""" ) def snake_case (__lowercase , __lowercase ) -> Union[str, Any]: '''simple docstring''' _snake_case : Dict = True _snake_case : Dict = flatten_dict(modela.params ) _snake_case : Any = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: _snake_case : Tuple = False return models_are_equal @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Tuple = BertConfig.from_pretrained("hf-internal-testing/tiny-bert-flax-only" ) _snake_case : List[Any] = FlaxBertModel(lowercase_ ) _snake_case : List[str] = "bert" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(lowercase_ , lowercase_ ) ) with self.assertRaises(lowercase_ ): _snake_case : Optional[Any] = FlaxBertModel.from_pretrained(lowercase_ ) _snake_case : Optional[int] = FlaxBertModel.from_pretrained(lowercase_ , subfolder=lowercase_ ) self.assertTrue(check_models_equal(lowercase_ , lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : Tuple = BertConfig.from_pretrained("hf-internal-testing/tiny-bert-flax-only" ) _snake_case : Optional[Any] = FlaxBertModel(lowercase_ ) _snake_case : List[Any] = "bert" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(lowercase_ , lowercase_ ) , max_shard_size="10KB" ) with self.assertRaises(lowercase_ ): _snake_case : List[Any] = FlaxBertModel.from_pretrained(lowercase_ ) _snake_case : Dict = FlaxBertModel.from_pretrained(lowercase_ , subfolder=lowercase_ ) self.assertTrue(check_models_equal(lowercase_ , lowercase_ ) ) def UpperCamelCase ( self ): _snake_case : Tuple = "bert" _snake_case : Dict = "hf-internal-testing/tiny-random-bert-subfolder" with self.assertRaises(lowercase_ ): _snake_case : Any = FlaxBertModel.from_pretrained(lowercase_ ) _snake_case : List[str] = FlaxBertModel.from_pretrained(lowercase_ , subfolder=lowercase_ ) self.assertIsNotNone(lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[str] = "bert" _snake_case : int = "hf-internal-testing/tiny-random-bert-sharded-subfolder" with self.assertRaises(lowercase_ ): _snake_case : Optional[int] = FlaxBertModel.from_pretrained(lowercase_ ) _snake_case : Dict = FlaxBertModel.from_pretrained(lowercase_ , subfolder=lowercase_ ) self.assertIsNotNone(lowercase_ )
670
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class lowercase_ : _lowerCamelCase = LEDConfig _lowerCamelCase = {} _lowerCamelCase = 'gelu' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=4 , ): _snake_case : Optional[int] = parent _snake_case : str = batch_size _snake_case : int = seq_length _snake_case : Dict = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = vocab_size _snake_case : str = hidden_size _snake_case : int = num_hidden_layers _snake_case : Union[str, Any] = num_attention_heads _snake_case : int = intermediate_size _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : int = max_position_embeddings _snake_case : Union[str, Any] = eos_token_id _snake_case : str = pad_token_id _snake_case : Any = bos_token_id _snake_case : str = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after _snake_case : List[Any] = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests _snake_case : List[str] = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def UpperCamelCase ( self ): _snake_case : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case : Any = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case : Optional[int] = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : List[str] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) _snake_case : Optional[Any] = prepare_led_inputs_dict(lowercase_ , lowercase_ , lowercase_ ) _snake_case : int = tf.concat( [tf.zeros_like(lowercase_ )[:, :-1], tf.ones_like(lowercase_ )[:, -1:]] , axis=-1 , ) _snake_case : List[Any] = global_attention_mask return config, inputs_dict def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Dict = TFLEDModel(config=lowercase_ ).get_decoder() _snake_case : Optional[Any] = inputs_dict["input_ids"] _snake_case : Optional[int] = input_ids[:1, :] _snake_case : int = inputs_dict["attention_mask"][:1, :] _snake_case : int = 1 # first forward pass _snake_case : str = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) _snake_case ,_snake_case : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case : Tuple = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case : List[str] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case : str = model(lowercase_ , attention_mask=lowercase_ )[0] _snake_case : List[str] = model(lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx] _snake_case : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase_ , lowercase_ , rtol=1e-3 ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , ) -> List[Any]: '''simple docstring''' if attention_mask is None: _snake_case : int = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case : Optional[int] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case : Tuple = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class lowercase_ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCamelCase = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () _lowerCamelCase = (TFLEDForConditionalGeneration,) if is_tf_available() else () _lowerCamelCase = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFLEDModelTester(self ) _snake_case : List[Any] = ConfigTester(self , config_class=lowercase_ ) def UpperCamelCase ( self ): self.config_tester.run_common_tests() def UpperCamelCase ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = tf.zeros_like(inputs_dict["attention_mask"] ) _snake_case : Tuple = 2 _snake_case : Dict = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) _snake_case : Tuple = True _snake_case : Union[str, Any] = self.model_tester.seq_length _snake_case : Union[str, Any] = self.model_tester.encoder_seq_length def check_decoder_attentions_output(lowercase_ ): _snake_case : Optional[Any] = outputs.decoder_attentions self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(lowercase_ ): _snake_case : int = [t.numpy() for t in outputs.encoder_attentions] _snake_case : Optional[int] = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(lowercase_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: _snake_case : Union[str, Any] = True _snake_case : Dict = False _snake_case : Any = False _snake_case : Any = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) _snake_case : Tuple = len(lowercase_ ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) if self.is_encoder_decoder: _snake_case : int = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_decoder_attentions_output(lowercase_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Any = model_class(lowercase_ ) _snake_case : Optional[Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) # Check attention is always last and order is fine _snake_case : Optional[int] = True _snake_case : Optional[int] = True _snake_case : List[Any] = model_class(lowercase_ ) _snake_case : Union[str, Any] = model(self._prepare_for_class(lowercase_ , lowercase_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowercase_ ) ) self.assertEqual(model.config.output_hidden_states , lowercase_ ) check_encoder_attentions_output(lowercase_ ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # TODO: Head-masking not yet implement pass def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' return tf.constant(__lowercase , dtype=tf.intaa ) __SCREAMING_SNAKE_CASE : List[Any] = 1E-4 @slow @require_tf class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Dict = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here _snake_case : Union[str, Any] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[int] = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Union[str, Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Optional[Any] = model(**lowercase_ )[0] _snake_case : str = (1, 1_024, 768) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[Any] = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 ) def UpperCamelCase ( self ): _snake_case : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here _snake_case : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : int = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) _snake_case : Optional[Any] = prepare_led_inputs_dict(model.config , lowercase_ , lowercase_ ) _snake_case : Tuple = model(**lowercase_ )[0] _snake_case : Any = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , lowercase_ ) # change to expected output here _snake_case : Optional[int] = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , lowercase_ , atol=1e-3 , rtol=1e-3 )
670
1
import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( 'The `image_to_image.py` script is outdated. Please use directly `from diffusers import' ' StableDiffusionImg2ImgPipeline` instead.' )
670
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ReformerTokenizer _lowerCamelCase = ReformerTokenizerFast _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True def UpperCamelCase ( self ): super().setUp() _snake_case : Union[str, Any] = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : int = "<s>" _snake_case : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def UpperCamelCase ( self ): _snake_case : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowercase_ ) , 1_000 ) def UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def UpperCamelCase ( self ): if not self.test_rust_tokenizer: return _snake_case : Tuple = self.get_tokenizer() _snake_case : List[str] = self.get_rust_tokenizer() _snake_case : int = "I was born in 92000, and this is falsé." _snake_case : Tuple = tokenizer.tokenize(lowercase_ ) _snake_case : List[Any] = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : str = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) _snake_case : Tuple = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) _snake_case : Dict = self.get_rust_tokenizer() _snake_case : List[Any] = tokenizer.encode(lowercase_ ) _snake_case : str = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def UpperCamelCase ( self , lowercase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(lowercase_ , **lowercase_ ) # Simple input _snake_case : List[str] = "This is a simple input" _snake_case : Optional[Any] = ["This is a simple input 1", "This is a simple input 2"] _snake_case : Union[str, Any] = ("This is a simple input", "This is a pair") _snake_case : int = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Simple input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises(lowercase_ , tokenizer_r.encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" ) # Pair input self.assertRaises( lowercase_ , tokenizer_r.batch_encode_plus , lowercase_ , max_length=lowercase_ , padding="max_length" , ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): _snake_case : Dict = ReformerTokenizer(lowercase_ , keep_accents=lowercase_ ) _snake_case : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [285, 46, 10, 170, 382] , ) _snake_case : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _snake_case : Any = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def UpperCamelCase ( self ): return ReformerTokenizer.from_pretrained("google/reformer-crime-and-punishment" ) @slow def UpperCamelCase ( self ): _snake_case : int = "Hello World!" _snake_case : Dict = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def UpperCamelCase ( self ): import torch from transformers import ReformerConfig, ReformerModel # Build sequence _snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] _snake_case : str = " ".join(lowercase_ ) _snake_case : Tuple = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" ) _snake_case : Tuple = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors="pt" ) _snake_case : int = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _snake_case : Union[str, Any] = encoded_sequence["input_ids"].shape _snake_case : List[str] = ReformerModel(lowercase_ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def UpperCamelCase ( self ): # fmt: off _snake_case : Union[str, Any] = {"input_ids": [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _snake_case : Tuple = [ "This is a very simple sentence.", "The quick brown fox jumps over the lazy dog.", ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/reformer-crime-and-punishment" , revision="0e6c3decb8211d49bf881013425dc8b0448b3f5a" , padding=lowercase_ , sequences=lowercase_ , )
670
1
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) def snake_case (__lowercase , __lowercase=False ) -> Optional[Any]: '''simple docstring''' _snake_case : List[Any] = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("head" ): _snake_case : Union[str, Any] = "segformer.encoder." + key if key.startswith("backbone" ): _snake_case : Optional[int] = key.replace("backbone" , "segformer.encoder" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _snake_case : str = key[key.find("patch_embed" ) + len("patch_embed" )] _snake_case : Dict = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(__lowercase )-1}""" ) if "norm" in key: _snake_case : str = key.replace("norm" , "layer_norm" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _snake_case : Optional[int] = key[key.find("segformer.encoder.layer_norm" ) + len("segformer.encoder.layer_norm" )] _snake_case : Optional[Any] = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(__lowercase )-1}""" ) if "layer_norm1" in key: _snake_case : str = key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: _snake_case : Any = key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 _snake_case : List[Any] = key[key.find("block" ) + len("block" )] _snake_case : Optional[Any] = key.replace(F"""block{idx}""" , F"""block.{int(__lowercase )-1}""" ) if "attn.q" in key: _snake_case : int = key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: _snake_case : Tuple = key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: _snake_case : List[str] = key.replace("attn" , "attention.self" ) if "fc1" in key: _snake_case : Optional[int] = key.replace("fc1" , "dense1" ) if "fc2" in key: _snake_case : List[str] = key.replace("fc2" , "dense2" ) if "linear_pred" in key: _snake_case : List[str] = key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: _snake_case : int = key.replace("linear_fuse.conv" , "linear_fuse" ) _snake_case : int = key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _snake_case : Any = key[key.find("linear_c" ) + len("linear_c" )] _snake_case : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(__lowercase )-1}""" ) if key.startswith("head" ): _snake_case : Optional[Any] = key.replace("head" , "classifier" ) _snake_case : int = value return new_state_dict def snake_case (__lowercase , __lowercase ) -> Dict: '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _snake_case : Optional[Any] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) _snake_case : Dict = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict _snake_case : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] _snake_case : List[str] = kv_bias[: config.hidden_sizes[i]] _snake_case : int = kv_weight[ config.hidden_sizes[i] :, : ] _snake_case : List[str] = kv_bias[ config.hidden_sizes[i] : ] def snake_case () -> Dict: '''simple docstring''' _snake_case : Any = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case : int = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return image @torch.no_grad() def snake_case (__lowercase , __lowercase , __lowercase ) -> Dict: '''simple docstring''' _snake_case : List[Any] = SegformerConfig() _snake_case : Dict = False # set attributes based on model_name _snake_case : Optional[int] = "huggingface/label-files" if "segformer" in model_name: _snake_case : Optional[Any] = model_name[len("segformer." ) : len("segformer." ) + 2] if "ade" in model_name: _snake_case : List[str] = 150 _snake_case : Any = "ade20k-id2label.json" _snake_case : Tuple = (1, 150, 128, 128) elif "city" in model_name: _snake_case : Any = 19 _snake_case : List[str] = "cityscapes-id2label.json" _snake_case : List[str] = (1, 19, 128, 128) else: raise ValueError(F"""Model {model_name} not supported""" ) elif "mit" in model_name: _snake_case : Dict = True _snake_case : List[str] = model_name[4:6] _snake_case : Union[str, Any] = 1_000 _snake_case : int = "imagenet-1k-id2label.json" _snake_case : List[str] = (1, 1_000) else: raise ValueError(F"""Model {model_name} not supported""" ) # set config attributes _snake_case : List[Any] = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type="dataset" ) , "r" ) ) _snake_case : Tuple = {int(__lowercase ): v for k, v in idalabel.items()} _snake_case : Union[str, Any] = idalabel _snake_case : List[Any] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _snake_case : str = [64, 128, 320, 512] _snake_case : Tuple = 256 elif size == "b2": _snake_case : Optional[Any] = [64, 128, 320, 512] _snake_case : int = 768 _snake_case : List[str] = [3, 4, 6, 3] elif size == "b3": _snake_case : List[str] = [64, 128, 320, 512] _snake_case : List[str] = 768 _snake_case : Tuple = [3, 4, 18, 3] elif size == "b4": _snake_case : int = [64, 128, 320, 512] _snake_case : Any = 768 _snake_case : Dict = [3, 8, 27, 3] elif size == "b5": _snake_case : List[Any] = [64, 128, 320, 512] _snake_case : List[Any] = 768 _snake_case : Dict = [3, 6, 40, 3] else: raise ValueError(F"""Size {size} not supported""" ) # load image processor (only resize + normalize) _snake_case : Any = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__lowercase , align=__lowercase , do_random_crop=__lowercase ) # prepare image _snake_case : Optional[Any] = prepare_img() _snake_case : Tuple = image_processor(images=__lowercase , return_tensors="pt" ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict if encoder_only: _snake_case : Dict = torch.load(__lowercase , map_location=torch.device("cpu" ) ) else: _snake_case : List[str] = torch.load(__lowercase , map_location=torch.device("cpu" ) )["state_dict"] # rename keys _snake_case : Optional[Any] = rename_keys(__lowercase , encoder_only=__lowercase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(__lowercase , __lowercase ) # create HuggingFace model and load state dict if encoder_only: _snake_case : Union[str, Any] = False _snake_case : Tuple = SegformerForImageClassification(__lowercase ) else: _snake_case : List[Any] = SegformerForSemanticSegmentation(__lowercase ) model.load_state_dict(__lowercase ) model.eval() # forward pass _snake_case : int = model(__lowercase ) _snake_case : Dict = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _snake_case : Dict = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _snake_case : Tuple = torch.tensor( [ [[-7.5820, -8.7231, -8.3215], [-8.0600, -10.3529, -10.0304], [-7.5208, -9.4103, -9.6239]], [[-12.6918, -13.8994, -13.7137], [-13.3196, -15.7523, -15.4789], [-12.9343, -14.8757, -14.9689]], [[-11.1911, -11.9421, -11.3243], [-11.3342, -13.6839, -13.3581], [-10.3909, -12.1832, -12.4858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _snake_case : str = torch.tensor( [ [[-11.8173, -14.3850, -16.3128], [-14.5648, -16.5804, -18.6568], [-14.7223, -15.7387, -18.4218]], [[-15.7290, -17.9171, -19.4423], [-18.3105, -19.9448, -21.4661], [-17.9296, -18.6497, -20.7910]], [[-15.0783, -17.0336, -18.2789], [-16.8771, -18.6870, -20.1612], [-16.2454, -17.1426, -19.5055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _snake_case : Optional[Any] = torch.tensor( [ [[-9.0878, -10.2081, -10.1891], [-9.3144, -10.7941, -10.9843], [-9.2294, -10.3855, -10.5704]], [[-12.2316, -13.9068, -13.6102], [-12.9161, -14.3702, -14.3235], [-12.5233, -13.7174, -13.7932]], [[-14.6275, -15.2490, -14.9727], [-14.3400, -15.9687, -16.2827], [-14.1484, -15.4033, -15.8937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _snake_case : Optional[Any] = torch.tensor( [ [[-12.3144, -13.2447, -14.0802], [-13.3614, -14.5816, -15.6117], [-13.3340, -14.4433, -16.2219]], [[-19.2781, -20.4128, -20.7506], [-20.6153, -21.6566, -22.0998], [-19.9800, -21.0430, -22.1494]], [[-18.8739, -19.7804, -21.1834], [-20.1233, -21.6765, -23.2944], [-20.0315, -21.2641, -23.6944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _snake_case : Optional[int] = torch.tensor( [ [[-9.5524, -12.0835, -11.7348], [-10.5229, -13.6446, -14.5662], [-9.5842, -12.8851, -13.9414]], [[-15.3432, -17.5323, -17.0818], [-16.3330, -18.9255, -19.2101], [-15.1340, -17.7848, -18.3971]], [[-12.6072, -14.9486, -14.6631], [-13.7629, -17.0907, -17.7745], [-12.7899, -16.1695, -17.1671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _snake_case : str = torch.tensor( [ [[-11.9295, -13.4057, -14.8106], [-13.3431, -14.8179, -15.3781], [-14.2836, -15.5942, -16.1588]], [[-11.4906, -12.8067, -13.6564], [-13.1189, -14.0500, -14.1543], [-13.8748, -14.5136, -14.8789]], [[0.5374, 0.1067, -0.4742], [0.1141, -0.2255, -0.7099], [-0.3000, -0.5924, -1.3105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _snake_case : Optional[int] = torch.tensor( [ [[-7.8217, -9.8767, -10.1717], [-9.4438, -10.9058, -11.4047], [-9.7939, -12.3495, -12.1079]], [[-7.1514, -9.5336, -10.0860], [-9.7776, -11.6822, -11.8439], [-10.1411, -12.7655, -12.8972]], [[0.3021, 0.0805, -0.2310], [-0.0328, -0.1605, -0.2714], [-0.1408, -0.5477, -0.6976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _snake_case : Tuple = torch.tensor( [ [ [-1.1372e01, -1.2787e01, -1.3477e01], [-1.2536e01, -1.4194e01, -1.4409e01], [-1.3217e01, -1.4888e01, -1.5327e01], ], [ [-1.4791e01, -1.7122e01, -1.8277e01], [-1.7163e01, -1.9192e01, -1.9533e01], [-1.7897e01, -1.9991e01, -2.0315e01], ], [ [7.6723e-01, 4.1921e-01, -7.7878e-02], [4.7772e-01, 9.5557e-03, -2.8082e-01], [3.6032e-01, -2.4826e-01, -5.1168e-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _snake_case : List[str] = torch.tensor( [ [[-9.4959, -11.3087, -11.7479], [-11.0025, -12.6540, -12.3319], [-11.4064, -13.0487, -12.9905]], [[-9.8905, -11.3084, -12.0854], [-11.1726, -12.7698, -12.9583], [-11.5985, -13.3278, -14.1774]], [[0.2213, 0.0192, -0.2466], [-0.1731, -0.4213, -0.4874], [-0.3126, -0.6541, -1.1389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _snake_case : List[str] = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _snake_case : Dict = torch.tensor( [ [[-16.0976, -16.4856, -17.3962], [-16.6234, -19.0342, -19.7685], [-16.0900, -18.0661, -19.1180]], [[-18.4750, -18.8488, -19.5074], [-19.4030, -22.1570, -22.5977], [-19.1191, -20.8486, -22.3783]], [[-4.5178, -5.5037, -6.5109], [-5.0884, -7.2174, -8.0334], [-4.4156, -5.8117, -7.2970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _snake_case : List[str] = torch.tensor( [ [[-14.2081, -14.4732, -14.1977], [-14.5867, -16.4423, -16.6356], [-13.4441, -14.9685, -16.8696]], [[-14.4576, -14.7073, -15.0451], [-15.0816, -17.6237, -17.9873], [-14.4213, -16.0199, -18.5992]], [[-4.7349, -4.9588, -5.0966], [-4.3210, -6.9325, -7.2591], [-3.4312, -4.7484, -7.1917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _snake_case : Optional[int] = torch.tensor( [ [[-11.7737, -11.9526, -11.3273], [-13.6692, -14.4574, -13.8878], [-13.8937, -14.6924, -15.9345]], [[-14.6706, -14.5330, -14.1306], [-16.1502, -16.8180, -16.4269], [-16.8338, -17.8939, -20.1746]], [[1.0491, 0.8289, 1.0310], [1.1044, 0.5219, 0.8055], [1.0899, 0.6926, 0.5590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _snake_case : Any = torch.tensor( [ [[-12.5641, -13.4777, -13.0684], [-13.9587, -15.8983, -16.6557], [-13.3109, -15.7350, -16.3141]], [[-14.7074, -15.4352, -14.5944], [-16.6353, -18.1663, -18.6120], [-15.1702, -18.0329, -18.1547]], [[-1.7990, -2.0951, -1.7784], [-2.6397, -3.8245, -3.9686], [-1.5264, -2.8126, -2.9316]], ] ) else: _snake_case : Dict = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , __lowercase , atol=1e-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(__lowercase ).mkdir(exist_ok=__lowercase ) model.save_pretrained(__lowercase ) image_processor.save_pretrained(__lowercase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() parser.add_argument( '--model_name', default='segformer.b0.512x512.ade.160k', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) __SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
670
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): _snake_case : Any = tempfile.mkdtemp() # fmt: off _snake_case : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _snake_case : Dict = dict(zip(lowercase_ , range(len(lowercase_ ) ) ) ) _snake_case : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] _snake_case : Optional[int] = {"unk_token": "<unk>"} _snake_case : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowercase_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowercase_ ) ) _snake_case : Any = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } _snake_case : Optional[Any] = os.path.join(self.tmpdirname , lowercase_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowercase_ , lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , **lowercase_ ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self ): _snake_case : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _snake_case : Union[str, Any] = [Image.fromarray(np.moveaxis(lowercase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self ): _snake_case : Tuple = self.get_tokenizer() _snake_case : Any = self.get_rust_tokenizer() _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_slow.save_pretrained(self.tmpdirname ) _snake_case : Optional[int] = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=lowercase_ ) _snake_case : List[Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) processor_fast.save_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowercase_ ) self.assertIsInstance(processor_fast.tokenizer , lowercase_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowercase_ ) self.assertIsInstance(processor_fast.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : List[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _snake_case : List[Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) _snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowercase_ , padding_value=1.0 ) _snake_case : Tuple = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowercase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : int = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Optional[Any] = image_processor(lowercase_ , return_tensors="np" ) _snake_case : str = processor(images=lowercase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = self.get_image_processor() _snake_case : Any = self.get_tokenizer() _snake_case : Dict = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[str] = "lower newer" _snake_case : int = processor(text=lowercase_ ) _snake_case : str = tokenizer(lowercase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self ): _snake_case : List[Any] = self.get_image_processor() _snake_case : int = self.get_tokenizer() _snake_case : Tuple = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : List[Any] = "lower newer" _snake_case : int = self.prepare_image_inputs() _snake_case : Dict = processor(text=lowercase_ , images=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[str] = self.get_tokenizer() _snake_case : Union[str, Any] = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = self.prepare_image_inputs() _snake_case : Dict = self.prepare_image_inputs() _snake_case : List[Any] = processor(images=lowercase_ , visual_prompt=lowercase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowercase_ ): processor() def UpperCamelCase ( self ): _snake_case : Dict = self.get_image_processor() _snake_case : List[Any] = self.get_tokenizer() _snake_case : str = CLIPSegProcessor(tokenizer=lowercase_ , image_processor=lowercase_ ) _snake_case : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _snake_case : Any = processor.batch_decode(lowercase_ ) _snake_case : Any = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ )
670
1
from typing import List from .keymap import KEYMAP, get_character def snake_case (__lowercase ) -> List[Any]: '''simple docstring''' def decorator(__lowercase ): _snake_case : List[str] = getattr(__lowercase , "handle_key" , [] ) handle += [key] setattr(__lowercase , "handle_key" , __lowercase ) return func return decorator def snake_case (*__lowercase ) -> Union[str, Any]: '''simple docstring''' def decorator(__lowercase ): _snake_case : Any = getattr(__lowercase , "handle_key" , [] ) handle += keys setattr(__lowercase , "handle_key" , __lowercase ) return func return decorator class lowercase_ ( __snake_case ): def __new__( cls , lowercase_ , lowercase_ , lowercase_ ): _snake_case : Dict = super().__new__(cls , lowercase_ , lowercase_ , lowercase_ ) if not hasattr(lowercase_ , "key_handler" ): setattr(lowercase_ , "key_handler" , {} ) setattr(lowercase_ , "handle_input" , KeyHandler.handle_input ) for value in attrs.values(): _snake_case : Optional[Any] = getattr(lowercase_ , "handle_key" , [] ) for key in handled_keys: _snake_case : Dict = value return new_cls @staticmethod def UpperCamelCase ( cls ): _snake_case : List[str] = get_character() if char != KEYMAP["undefined"]: _snake_case : str = ord(lowercase_ ) _snake_case : List[Any] = cls.key_handler.get(lowercase_ ) if handler: _snake_case : int = char return handler(cls ) else: return None def snake_case (cls ) -> Any: '''simple docstring''' return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
670
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case (__lowercase ) -> Any: '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(__lowercase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def snake_case (__lowercase ) -> Any: '''simple docstring''' _snake_case : int = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) _snake_case : List[Any] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format _snake_case : Optional[int] = PipelineDataFormat.from_str( format=__lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(__lowercase , __lowercase ) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ , lowercase_ ): _snake_case : str = nlp _snake_case : str = reader @staticmethod def UpperCamelCase ( lowercase_ ): _snake_case : Dict = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=lowercase_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=lowercase_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=lowercase_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=lowercase_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=lowercase_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=lowercase_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=lowercase_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=lowercase_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=lowercase_ ) def UpperCamelCase ( self ): _snake_case ,_snake_case : Tuple = self._nlp, [] for entry in self._reader: _snake_case : Optional[Any] = nlp(**lowercase_ ) if self._reader.is_multi_columns else nlp(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): outputs.append(lowercase_ ) else: outputs += output # Saving data if self._nlp.binary_output: _snake_case : str = self._reader.save_binary(lowercase_ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(lowercase_ )
670
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'encoder-decoder' _lowerCamelCase = True def __init__( self , **lowercase_ ): super().__init__(**lowercase_ ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" _snake_case : int = kwargs.pop("encoder" ) _snake_case : Tuple = encoder_config.pop("model_type" ) _snake_case : Dict = kwargs.pop("decoder" ) _snake_case : Optional[Any] = decoder_config.pop("model_type" ) from ..auto.configuration_auto import AutoConfig _snake_case : Tuple = AutoConfig.for_model(lowercase_ , **lowercase_ ) _snake_case : Optional[Any] = AutoConfig.for_model(lowercase_ , **lowercase_ ) _snake_case : Tuple = True @classmethod def UpperCamelCase ( cls , lowercase_ , lowercase_ , **lowercase_ ): logger.info("Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config" ) _snake_case : List[Any] = True _snake_case : str = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **lowercase_ ) def UpperCamelCase ( self ): _snake_case : Union[str, Any] = copy.deepcopy(self.__dict__ ) _snake_case : Dict = self.encoder.to_dict() _snake_case : Dict = self.decoder.to_dict() _snake_case : int = self.__class__.model_type return output
670
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): def __init__( self , lowercase_ ): super().__init__() _snake_case : List[str] = nn.ModuleList(lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = True , ): for i, (image, scale, controlnet) in enumerate(zip(lowercase_ , lowercase_ , self.nets ) ): _snake_case ,_snake_case : Optional[int] = controlnet( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ) # merge samples if i == 0: _snake_case ,_snake_case : Tuple = down_samples, mid_sample else: _snake_case : Tuple = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase_ , lowercase_ ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def UpperCamelCase ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , ): _snake_case : Tuple = 0 _snake_case : Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase_ , is_main_process=lowercase_ , save_function=lowercase_ , safe_serialization=lowercase_ , variant=lowercase_ , ) idx += 1 _snake_case : int = model_path_to_save + f"""_{idx}""" @classmethod def UpperCamelCase ( cls , lowercase_ , **lowercase_ ): _snake_case : List[str] = 0 _snake_case : Optional[Any] = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case : Optional[Any] = pretrained_model_path while os.path.isdir(lowercase_ ): _snake_case : int = ControlNetModel.from_pretrained(lowercase_ , **lowercase_ ) controlnets.append(lowercase_ ) idx += 1 _snake_case : str = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(lowercase_ )} controlnets loaded from {pretrained_model_path}.""" ) if len(lowercase_ ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(lowercase_ )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(lowercase_ )
670
1
import math import random def snake_case (__lowercase , __lowercase = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __SCREAMING_SNAKE_CASE : int = 0.02 def snake_case (__lowercase , __lowercase ) -> float: '''simple docstring''' _snake_case : List[Any] = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__lowercase ): # Forward propagation _snake_case : int = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? _snake_case : int = (expected / 100) - layer_a # Error delta _snake_case : str = layer_1_error * sigmoid_function(__lowercase , __lowercase ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : Union[str, Any] = int(input('Expected value: ')) __SCREAMING_SNAKE_CASE : Optional[Any] = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
670
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowercase_ ( __snake_case ): _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'CLIPImageProcessor' _lowerCamelCase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_ ): _snake_case : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowercase_ , ) _snake_case : Dict = kwargs.pop("feature_extractor" ) _snake_case : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowercase_ , lowercase_ ) def __call__( self , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _snake_case : str = self.tokenizer(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if images is not None: _snake_case : List[str] = self.image_processor(lowercase_ , return_tensors=lowercase_ , **lowercase_ ) if text is not None and images is not None: _snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase_ ) , tensor_type=lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.batch_decode(*lowercase_ , **lowercase_ ) def UpperCamelCase ( self , *lowercase_ , **lowercase_ ): return self.tokenizer.decode(*lowercase_ , **lowercase_ ) @property def UpperCamelCase ( self ): _snake_case : Any = self.tokenizer.model_input_names _snake_case : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
670
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Tuple = { 'configuration_swinv2': ['SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Swinv2Config'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ 'SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'Swinv2ForImageClassification', 'Swinv2ForMaskedImageModeling', 'Swinv2Model', 'Swinv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
from __future__ import annotations def snake_case (__lowercase , __lowercase , __lowercase ) -> dict[str, float]: '''simple docstring''' if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
670
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Tuple = { 'configuration_mgp_str': ['MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MgpstrConfig'], 'processing_mgp_str': ['MgpstrProcessor'], 'tokenization_mgp_str': ['MgpstrTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ 'MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST', 'MgpstrModel', 'MgpstrPreTrainedModel', 'MgpstrForSceneTextRecognition', ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def snake_case (*__lowercase ) -> Dict: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): _snake_case : Dict = list(__lowercase ) for i in range(len(__lowercase ) ): _snake_case : List[str] = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : str = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowercase , __lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def snake_case (__lowercase = None , __lowercase = 128 ) -> Any: '''simple docstring''' if function is None: return functools.partial(__lowercase , starting_batch_size=__lowercase ) _snake_case : List[str] = starting_batch_size def decorator(*__lowercase , **__lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _snake_case : Optional[Any] = list(inspect.signature(__lowercase ).parameters.keys() ) # Guard against user error if len(__lowercase ) < (len(__lowercase ) + 1): _snake_case : str = ", ".join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowercase , *__lowercase , **__lowercase ) except Exception as e: if should_reduce_batch_size(__lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
670
1
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument __SCREAMING_SNAKE_CASE : Union[str, Any] = { '/attention/': '/0/SelfAttention/', '/self_attention/': '/0/SelfAttention/', '/encoder_decoder_attention/': '/1/EncDecAttention/', 'value': 'v', 'query': 'q', 'key': 'k', 'out': 'o', 'pre_self_attention_layer_norm': '0/layer_norm', 'pre_cross_attention_layer_norm': '1/layer_norm', 'pre_attention_layer_norm': '0/layer_norm', # previously 1, but seems wrong 'token_embedder': 'shared', 'encoder_norm': 'final_layer_norm', 'decoder_norm': 'final_layer_norm', 'relpos_bias/rel_embedding': 'block/0/layer/0/SelfAttention/relative_attention_bias/weight', 'router/router_weights/w/': 'router/classifier/', 'roer/roer_weights/w/': 'router/classifier/', 'logits_dense': 'lm_head', } def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case : str = list(s_dict.keys() ) for key in keys: _snake_case : Any = r".*/layers_(\d+)" _snake_case : str = key if re.match(__lowercase , __lowercase ): _snake_case : Optional[Any] = re.sub(r"layers_(\d+)" , r"block/\1/layer" , __lowercase ) _snake_case : Union[str, Any] = r"(encoder|decoder)\/" if re.match(__lowercase , __lowercase ): _snake_case : Optional[int] = re.match(__lowercase , __lowercase ).groups() if groups[0] == "encoder": _snake_case : Any = re.sub(r"/mlp/" , r"/1/mlp/" , __lowercase ) _snake_case : Optional[int] = re.sub(r"/pre_mlp_layer_norm/" , r"/1/layer_norm/" , __lowercase ) elif groups[0] == "decoder": _snake_case : Union[str, Any] = re.sub(r"/mlp/" , r"/2/mlp/" , __lowercase ) _snake_case : str = re.sub(r"/pre_mlp_layer_norm/" , r"/2/layer_norm/" , __lowercase ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: _snake_case : Optional[int] = new_key.replace(__lowercase , __lowercase ) print(F"""{key} -> {new_key}""" ) _snake_case : Tuple = s_dict.pop(__lowercase ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: _snake_case : List[str] = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: _snake_case : str = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: _snake_case : List[str] = s_dict[key].shape[0] _snake_case : Optional[int] = s_dict[key] for idx in range(__lowercase ): _snake_case : List[Any] = expert_weihts[idx] print(F"""{key} -> {key.replace('expert/' , 'nested fstring' )}""" ) s_dict.pop(__lowercase ) return s_dict __SCREAMING_SNAKE_CASE : str = { 'NUM_ENCODER_LAYERS': 'num_layers', 'NUM_DECODER_LAYERS': 'num_decoder_layers', 'NUM_HEADS': 'num_heads', 'HEAD_DIM': 'd_kv', 'EMBED_DIM': 'd_model', 'MLP_DIM': 'd_ff', 'NUM_SELECTED_EXPERTS': 'num_selected_experts', 'NUM_ENCODER_SPARSE_LAYERS': 'num_sparse_encoder_layers', 'NUM_DECODER_SPARSE_LAYERS': 'num_sparse_decoder_layers', 'dense.MlpBlock.activations': 'feed_forward_proj', } def snake_case (__lowercase , __lowercase ) -> Any: '''simple docstring''' import regex as re with open(__lowercase , "r" ) as f: _snake_case : List[str] = f.read() _snake_case : Tuple = re.findall(r"(.*) = ([0-9.]*)" , __lowercase ) _snake_case : int = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": _snake_case : Dict = float(__lowercase ) if "." in value else int(__lowercase ) _snake_case : Tuple = re.findall(r"(.*activations) = \(\'(.*)\',\)" , __lowercase )[0] _snake_case : Dict = str(activation[1] ) _snake_case : Any = num_experts _snake_case : int = SwitchTransformersConfig(**__lowercase ) return config def snake_case (__lowercase , __lowercase , __lowercase=None , __lowercase="./" , __lowercase=8 ) -> str: '''simple docstring''' print(F"""Loading flax weights from : {flax_checkpoint_path}""" ) _snake_case : Optional[int] = checkpoints.load_tax_checkpoint(__lowercase ) if gin_file is not None: _snake_case : List[Any] = convert_gin_to_config(__lowercase , __lowercase ) else: _snake_case : Optional[Any] = SwitchTransformersConfig.from_pretrained(__lowercase ) _snake_case : Tuple = SwitchTransformersForConditionalGeneration(__lowercase ) _snake_case : List[Any] = flax_params["target"] _snake_case : str = flatten_dict(__lowercase , sep="/" ) _snake_case : str = rename_keys(__lowercase ) _snake_case : str = unflatten_dict(__lowercase , sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(__lowercase , __lowercase ) print(F"""Save PyTorch model to {pytorch_dump_path}""" ) pt_model.save_pretrained(__lowercase ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the' ' model architecture. If not provided, a `gin_file` has to be provided.' ), ) parser.add_argument( '--gin_file', default=None, type=str, required=False, help='Path to the gin config file. If not provided, a `config_file` has to be passed ', ) parser.add_argument( '--config_name', default=None, type=str, required=False, help='Config name of SwitchTransformers model.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output pytorch model.' ) parser.add_argument('--num_experts', default=8, type=int, required=False, help='Number of experts') __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
670
__SCREAMING_SNAKE_CASE : Union[str, Any] = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', 'p': 'ABBBA', 'q': 'ABBBB', 'r': 'BAAAA', 's': 'BAAAB', 't': 'BAABA', 'u': 'BAABB', 'v': 'BBBAB', 'w': 'BABAA', 'x': 'BABAB', 'y': 'BABBA', 'z': 'BABBB', ' ': ' ', } __SCREAMING_SNAKE_CASE : int = {value: key for key, value in encode_dict.items()} def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : Any = "" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("encode() accepts only letters of the alphabet and spaces" ) return encoded def snake_case (__lowercase ) -> str: '''simple docstring''' if set(__lowercase ) - {"A", "B", " "} != set(): raise Exception("decode() accepts only 'A', 'B' and spaces" ) _snake_case : str = "" for word in coded.split(): while len(__lowercase ) != 0: decoded += decode_dict[word[:5]] _snake_case : int = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
670
1
import argparse import os import re __SCREAMING_SNAKE_CASE : Tuple = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __SCREAMING_SNAKE_CASE : List[Any] = re.compile(R'[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict') # re pattern that matches identifiers in mappings __SCREAMING_SNAKE_CASE : Tuple = re.compile(R'\s*\(\s*"(\S[^"]+)"') def snake_case (__lowercase , __lowercase = False ) -> List[str]: '''simple docstring''' with open(__lowercase , "r" , encoding="utf-8" ) as f: _snake_case : Optional[Any] = f.read() _snake_case : List[str] = content.split("\n" ) _snake_case : Tuple = [] _snake_case : Optional[int] = 0 while line_idx < len(__lowercase ): if _re_intro_mapping.search(lines[line_idx] ) is not None: _snake_case : List[Any] = len(re.search(r"^(\s*)\S" , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(" " * indent + "(" ): new_lines.append(lines[line_idx] ) line_idx += 1 _snake_case : Any = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": _snake_case : Union[str, Any] = line_idx while not lines[line_idx].startswith(" " * indent + ")" ): line_idx += 1 blocks.append("\n".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers _snake_case : Dict = sorted(__lowercase , key=lambda __lowercase : _re_identifier.search(__lowercase ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(__lowercase , "w" , encoding="utf-8" ) as f: f.write("\n".join(__lowercase ) ) elif "\n".join(__lowercase ) != content: return True def snake_case (__lowercase = False ) -> Union[str, Any]: '''simple docstring''' _snake_case : Dict = [os.path.join(__lowercase , __lowercase ) for f in os.listdir(__lowercase ) if f.endswith(".py" )] _snake_case : Optional[Any] = [sort_auto_mapping(__lowercase , overwrite=__lowercase ) for fname in fnames] if not overwrite and any(__lowercase ): _snake_case : Optional[int] = [f for f, d in zip(__lowercase , __lowercase ) if d] raise ValueError( F"""The following files have auto mappings that need sorting: {', '.join(__lowercase )}. Run `make style` to fix""" " this." ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') __SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() sort_all_auto_mappings(not args.check_only)
670
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase ( self ): _snake_case ,_snake_case : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : List[Any] = "A painting of a squirrel eating a burger" _snake_case : Union[str, Any] = jax.device_count() _snake_case : List[Any] = num_samples * [prompt] _snake_case : Tuple = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : str = replicate(lowercase_ ) _snake_case : Dict = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : List[Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : Tuple = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[Any] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : str = images[0, 253:256, 253:256, -1] _snake_case : Tuple = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Optional[Any] = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ): _snake_case : Optional[Any] = "stabilityai/stable-diffusion-2" _snake_case ,_snake_case : List[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(lowercase_ , subfolder="scheduler" ) _snake_case ,_snake_case : int = FlaxStableDiffusionPipeline.from_pretrained( lowercase_ , scheduler=lowercase_ , revision="bf16" , dtype=jnp.bfloataa , ) _snake_case : str = scheduler_params _snake_case : Dict = "A painting of a squirrel eating a burger" _snake_case : Dict = jax.device_count() _snake_case : Optional[int] = num_samples * [prompt] _snake_case : List[str] = sd_pipe.prepare_inputs(lowercase_ ) _snake_case : Optional[int] = replicate(lowercase_ ) _snake_case : Union[str, Any] = shard(lowercase_ ) _snake_case : List[Any] = jax.random.PRNGKey(0 ) _snake_case : Union[str, Any] = jax.random.split(lowercase_ , jax.device_count() ) _snake_case : str = sd_pipe(lowercase_ , lowercase_ , lowercase_ , num_inference_steps=25 , jit=lowercase_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _snake_case : List[str] = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _snake_case : List[str] = images[0, 253:256, 253:256, -1] _snake_case : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _snake_case : Dict = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(f"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
670
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __SCREAMING_SNAKE_CASE : Any = { 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = ['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[str] = ['LayoutLMv2FeatureExtractor'] __SCREAMING_SNAKE_CASE : List[str] = ['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
670
from manim import * class lowercase_ ( __snake_case ): def UpperCamelCase ( self ): _snake_case : Tuple = Rectangle(height=0.5 , width=0.5 ) _snake_case : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _snake_case : List[str] = [mem.copy() for i in range(6 )] _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = VGroup(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : int = Text("CPU" , font_size=24 ) _snake_case : str = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) _snake_case : int = [mem.copy() for i in range(4 )] _snake_case : Dict = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : str = Text("GPU" , font_size=24 ) _snake_case : Optional[int] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) _snake_case : Any = [mem.copy() for i in range(6 )] _snake_case : Any = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Dict = Text("Model" , font_size=24 ) _snake_case : Dict = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , buff=0.5 , aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) _snake_case : str = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case : Union[str, Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowercase_ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowercase_ , buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) _snake_case : List[Any] = [mem.copy() for i in range(6 )] _snake_case : Union[str, Any] = VGroup(*lowercase_ ).arrange(lowercase_ , buff=0 ) _snake_case : Optional[Any] = Text("Loaded Checkpoint" , font_size=24 ) _snake_case : Union[str, Any] = Group(lowercase_ , lowercase_ ).arrange(lowercase_ , aligned_edge=lowercase_ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case : Optional[Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ , lowercase_ ) _snake_case : Union[str, Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowercase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _snake_case : List[Any] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) , Write(lowercase_ ) ) self.play(Write(lowercase_ , run_time=1 ) , Create(lowercase_ , run_time=1 ) ) _snake_case : int = [] _snake_case : str = [] for i, rect in enumerate(lowercase_ ): _snake_case : Dict = fill.copy().set_fill(lowercase_ , opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ , run_time=1 ) ) _snake_case : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ , run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
670
1
import re import string import numpy as np import datasets __SCREAMING_SNAKE_CASE : Dict = '\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n' __SCREAMING_SNAKE_CASE : int = '\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results["exact_match"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."]\n >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 33.3\n\n' __SCREAMING_SNAKE_CASE : Optional[Any] = '\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ ( datasets.Metric ): def UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , reference_urls=[] , ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=False , lowercase_=False , lowercase_=False , ): if regexes_to_ignore is not None: for s in regexes_to_ignore: _snake_case : Dict = np.array([re.sub(lowercase_ , "" , lowercase_ ) for x in predictions] ) _snake_case : Union[str, Any] = np.array([re.sub(lowercase_ , "" , lowercase_ ) for x in references] ) else: _snake_case : Any = np.asarray(lowercase_ ) _snake_case : Optional[int] = np.asarray(lowercase_ ) if ignore_case: _snake_case : Optional[Any] = np.char.lower(lowercase_ ) _snake_case : Dict = np.char.lower(lowercase_ ) if ignore_punctuation: _snake_case : Union[str, Any] = string.punctuation.maketrans("" , "" , string.punctuation ) _snake_case : Optional[Any] = np.char.translate(lowercase_ , table=lowercase_ ) _snake_case : List[Any] = np.char.translate(lowercase_ , table=lowercase_ ) if ignore_numbers: _snake_case : Tuple = string.digits.maketrans("" , "" , string.digits ) _snake_case : int = np.char.translate(lowercase_ , table=lowercase_ ) _snake_case : Optional[Any] = np.char.translate(lowercase_ , table=lowercase_ ) _snake_case : Dict = predictions == references return {"exact_match": np.mean(lowercase_ ) * 100}
670
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = 'linear' _lowerCamelCase = 'cosine' _lowerCamelCase = 'cosine_with_restarts' _lowerCamelCase = 'polynomial' _lowerCamelCase = 'constant' _lowerCamelCase = 'constant_with_warmup' _lowerCamelCase = 'piecewise_constant' def snake_case (__lowercase , __lowercase = -1 ) -> List[Any]: '''simple docstring''' return LambdaLR(__lowercase , lambda __lowercase : 1 , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1.0 , __lowercase ) ) return 1.0 return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' _snake_case : Optional[Any] = {} _snake_case : Optional[int] = step_rules.split("," ) for rule_str in rule_list[:-1]: _snake_case ,_snake_case : str = rule_str.split(":" ) _snake_case : Dict = int(__lowercase ) _snake_case : List[str] = float(__lowercase ) _snake_case : Tuple = value _snake_case : str = float(rule_list[-1] ) def create_rules_function(__lowercase , __lowercase ): def rule_func(__lowercase ) -> float: _snake_case : List[str] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowercase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _snake_case : int = create_rules_function(__lowercase , __lowercase ) return LambdaLR(__lowercase , __lowercase , last_epoch=__lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=-1 ) -> List[str]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 0.5 , __lowercase = -1 ) -> Dict: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Optional[int] = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowercase ) * 2.0 * progress )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase = 1 , __lowercase = -1 ) -> Optional[int]: '''simple docstring''' def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) _snake_case : Any = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowercase ) * progress) % 1.0) )) ) return LambdaLR(__lowercase , __lowercase , __lowercase ) def snake_case (__lowercase , __lowercase , __lowercase , __lowercase=1e-7 , __lowercase=1.0 , __lowercase=-1 ) -> List[Any]: '''simple docstring''' _snake_case : List[Any] = optimizer.defaults["lr"] if not (lr_init > lr_end): raise ValueError(F"""lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})""" ) def lr_lambda(__lowercase ): if current_step < num_warmup_steps: return float(__lowercase ) / float(max(1 , __lowercase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _snake_case : Tuple = lr_init - lr_end _snake_case : Any = num_training_steps - num_warmup_steps _snake_case : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps _snake_case : Optional[Any] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowercase , __lowercase , __lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case (__lowercase , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = 1 , __lowercase = 1.0 , __lowercase = -1 , ) -> List[Any]: '''simple docstring''' _snake_case : Any = SchedulerType(__lowercase ) _snake_case : Union[str, Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowercase , last_epoch=__lowercase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowercase , step_rules=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F"""{name} requires `num_warmup_steps`, please provide that argument.""" ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowercase , num_warmup_steps=__lowercase , last_epoch=__lowercase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F"""{name} requires `num_training_steps`, please provide that argument.""" ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , num_cycles=__lowercase , last_epoch=__lowercase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , power=__lowercase , last_epoch=__lowercase , ) return schedule_func( __lowercase , num_warmup_steps=__lowercase , num_training_steps=__lowercase , last_epoch=__lowercase )
670
1