code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
lowerCAmelCase__ : Any = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
143
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : str=1_3 , SCREAMING_SNAKE_CASE__ : Optional[int]=7 , SCREAMING_SNAKE_CASE__ : str=True , SCREAMING_SNAKE_CASE__ : int=True , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : str=True , SCREAMING_SNAKE_CASE__ : str=9_9 , SCREAMING_SNAKE_CASE__ : str=2_4 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=2 , SCREAMING_SNAKE_CASE__ : Optional[Any]=6 , SCREAMING_SNAKE_CASE__ : Optional[int]=3_7 , SCREAMING_SNAKE_CASE__ : List[Any]="gelu" , SCREAMING_SNAKE_CASE__ : str=0.1 , SCREAMING_SNAKE_CASE__ : List[Any]=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=5_1_2 , SCREAMING_SNAKE_CASE__ : List[str]=1_6 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=2 , SCREAMING_SNAKE_CASE__ : int=0.02 , SCREAMING_SNAKE_CASE__ : Optional[Any]=3 , SCREAMING_SNAKE_CASE__ : Optional[int]=None , SCREAMING_SNAKE_CASE__ : Tuple=1_0_0_0 , ) -> str: a_ : Optional[Any] = parent a_ : List[str] = batch_size a_ : List[str] = seq_length a_ : str = is_training a_ : str = use_input_mask a_ : int = use_token_type_ids a_ : List[str] = use_labels a_ : Optional[int] = vocab_size a_ : Any = hidden_size a_ : int = num_hidden_layers a_ : List[str] = num_attention_heads a_ : str = intermediate_size a_ : Union[str, Any] = hidden_act a_ : List[str] = hidden_dropout_prob a_ : int = attention_probs_dropout_prob a_ : int = max_position_embeddings a_ : Tuple = type_vocab_size a_ : Optional[Any] = type_sequence_label_size a_ : Tuple = initializer_range a_ : Dict = num_labels a_ : str = scope a_ : Optional[int] = range_bbox def SCREAMING_SNAKE_CASE ( self : List[str] ) -> int: a_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a_ : Any = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: a_ : int = bbox[i, j, 3] a_ : str = bbox[i, j, 1] a_ : List[str] = t if bbox[i, j, 2] < bbox[i, j, 0]: a_ : Tuple = bbox[i, j, 2] a_ : List[str] = bbox[i, j, 0] a_ : Union[str, Any] = t a_ : List[Any] = None if self.use_input_mask: a_ : Dict = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) a_ : List[Any] = None if self.use_token_type_ids: a_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a_ : int = None a_ : Tuple = None if self.use_labels: a_ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a_ : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a_ : Optional[int] = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , ) -> str: a_ : Any = LiltModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : Any = model(SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = model(SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = model(SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Tuple , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[Any] , ) -> int: a_ : Any = self.num_labels a_ : str = LiltForTokenClassification(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : str = model( SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Any , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] , ) -> str: a_ : Union[str, Any] = LiltForQuestionAnswering(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : List[str] = model( SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ , attention_mask=SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , start_positions=SCREAMING_SNAKE_CASE__ , end_positions=SCREAMING_SNAKE_CASE__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: a_ : int = self.prepare_config_and_inputs() ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) : List[Any] = config_and_inputs a_ : Optional[int] = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( lowercase__ , lowercase__ , lowercase__ , unittest.TestCase ): snake_case__ : Union[str, Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) snake_case__ : str = ( { '''feature-extraction''': LiltModel, '''question-answering''': LiltForQuestionAnswering, '''text-classification''': LiltForSequenceClassification, '''token-classification''': LiltForTokenClassification, '''zero-shot''': LiltForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : List[str] = False snake_case__ : str = False def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int ) -> int: return True def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Tuple: a_ : str = LiltModelTester(self ) a_ : List[Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> List[Any]: self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> str: a_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[int]: a_ : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: a_ : List[str] = type self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : int ) -> Optional[Any]: a_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[str]: a_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ : List[Any] = LiltModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @require_torch @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Union[str, Any]: a_ : List[str] = LiltModel.from_pretrained('SCUT-DLVCLab/lilt-roberta-en-base' ).to(SCREAMING_SNAKE_CASE__ ) a_ : str = torch.tensor([[1, 2]] , device=SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): a_ : str = model(input_ids=SCREAMING_SNAKE_CASE__ , bbox=SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = torch.Size([1, 2, 7_6_8] ) a_ : int = torch.tensor( [[-0.0653, 0.0950, -0.0061], [-0.0545, 0.0926, -0.0324]] , device=SCREAMING_SNAKE_CASE__ , ) self.assertTrue(outputs.last_hidden_state.shape , SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-3 ) )
32
0
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def a ( _UpperCAmelCase : Optional[int] ): '''simple docstring''' if "img_encoder.pos_embed" in name: __UpperCAmelCase : str = name.replace('''img_encoder.pos_embed''' , '''vision_model.embeddings.position_embeddings''' ) if "img_encoder.patch_embed.proj" in name: __UpperCAmelCase : Tuple = name.replace('''img_encoder.patch_embed.proj''' , '''vision_model.embeddings.patch_embeddings.projection''' ) if "img_encoder.patch_embed.norm" in name: __UpperCAmelCase : Optional[Any] = name.replace('''img_encoder.patch_embed.norm''' , '''vision_model.embeddings.layernorm''' ) if "img_encoder.layers" in name: __UpperCAmelCase : Optional[int] = name.replace('''img_encoder.layers''' , '''vision_model.encoder.stages''' ) if "blocks" in name and "res" not in name: __UpperCAmelCase : Union[str, Any] = name.replace('''blocks''' , '''layers''' ) if "attn" in name and "pre_assign" not in name: __UpperCAmelCase : int = name.replace('''attn''' , '''self_attn''' ) if "proj" in name and "self_attn" in name and "text" not in name: __UpperCAmelCase : Optional[Any] = name.replace('''proj''' , '''out_proj''' ) if "pre_assign_attn.attn.proj" in name: __UpperCAmelCase : str = name.replace('''pre_assign_attn.attn.proj''' , '''pre_assign_attn.attn.out_proj''' ) if "norm1" in name: __UpperCAmelCase : List[Any] = name.replace('''norm1''' , '''layer_norm1''' ) if "norm2" in name and "pre_assign" not in name: __UpperCAmelCase : str = name.replace('''norm2''' , '''layer_norm2''' ) if "img_encoder.norm" in name: __UpperCAmelCase : Optional[Any] = name.replace('''img_encoder.norm''' , '''vision_model.layernorm''' ) # text encoder if "text_encoder.token_embedding" in name: __UpperCAmelCase : List[Any] = name.replace('''text_encoder.token_embedding''' , '''text_model.embeddings.token_embedding''' ) if "text_encoder.positional_embedding" in name: __UpperCAmelCase : Optional[Any] = name.replace('''text_encoder.positional_embedding''' , '''text_model.embeddings.position_embedding.weight''' ) if "text_encoder.transformer.resblocks." in name: __UpperCAmelCase : List[Any] = name.replace('''text_encoder.transformer.resblocks.''' , '''text_model.encoder.layers.''' ) if "ln_1" in name: __UpperCAmelCase : Tuple = name.replace('''ln_1''' , '''layer_norm1''' ) if "ln_2" in name: __UpperCAmelCase : Any = name.replace('''ln_2''' , '''layer_norm2''' ) if "c_fc" in name: __UpperCAmelCase : Any = name.replace('''c_fc''' , '''fc1''' ) if "c_proj" in name: __UpperCAmelCase : Any = name.replace('''c_proj''' , '''fc2''' ) if "text_encoder" in name: __UpperCAmelCase : Optional[int] = name.replace('''text_encoder''' , '''text_model''' ) if "ln_final" in name: __UpperCAmelCase : int = name.replace('''ln_final''' , '''final_layer_norm''' ) # projection layers if "img_projector.linear_hidden." in name: __UpperCAmelCase : Union[str, Any] = name.replace('''img_projector.linear_hidden.''' , '''visual_projection.''' ) if "img_projector.linear_out." in name: __UpperCAmelCase : Tuple = name.replace('''img_projector.linear_out.''' , '''visual_projection.3.''' ) if "text_projector.linear_hidden" in name: __UpperCAmelCase : List[str] = name.replace('''text_projector.linear_hidden''' , '''text_projection''' ) if "text_projector.linear_out" in name: __UpperCAmelCase : Optional[Any] = name.replace('''text_projector.linear_out''' , '''text_projection.3''' ) return name def a ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] ): '''simple docstring''' for key in orig_state_dict.copy().keys(): __UpperCAmelCase : str = orig_state_dict.pop(__A ) if "qkv" in key: # weights and biases of the key, value and query projections of vision encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors __UpperCAmelCase : Optional[int] = key.split('''.''' ) __UpperCAmelCase : List[str] = int(key_split[2] ), int(key_split[4] ) __UpperCAmelCase : str = config.vision_config.hidden_size if "weight" in key: __UpperCAmelCase : Dict = val[:dim, :] __UpperCAmelCase : List[Any] = val[dim : dim * 2, :] __UpperCAmelCase : str = val[-dim:, :] else: __UpperCAmelCase : List[str] = val[:dim] __UpperCAmelCase : Union[str, Any] = val[dim : dim * 2] __UpperCAmelCase : Any = val[-dim:] elif "in_proj" in key: # weights and biases of the key, value and query projections of text encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors __UpperCAmelCase : Optional[int] = key.split('''.''' ) __UpperCAmelCase : Optional[Any] = int(key_split[3] ) __UpperCAmelCase : List[str] = config.text_config.hidden_size if "weight" in key: __UpperCAmelCase : int = val[:dim, :] __UpperCAmelCase : str = val[ dim : dim * 2, : ] __UpperCAmelCase : List[Any] = val[-dim:, :] else: __UpperCAmelCase : str = val[:dim] __UpperCAmelCase : Union[str, Any] = val[dim : dim * 2] __UpperCAmelCase : Dict = val[-dim:] else: __UpperCAmelCase : List[Any] = rename_key(__A ) # squeeze if necessary if ( "text_projection.0" in new_name or "text_projection.3" in new_name or "visual_projection.0" in new_name or "visual_projection.3" in new_name ): __UpperCAmelCase : Tuple = val.squeeze_() else: __UpperCAmelCase : List[str] = val return orig_state_dict def a ( ): '''simple docstring''' __UpperCAmelCase : List[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' __UpperCAmelCase : List[str] = Image.open(requests.get(__A , stream=__A ).raw ) return im @torch.no_grad() def a ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any]="groupvit-gcc-yfcc" , _UpperCAmelCase : Optional[Any]=False ): '''simple docstring''' __UpperCAmelCase : Optional[int] = GroupViTConfig() __UpperCAmelCase : Union[str, Any] = GroupViTModel(__A ).eval() __UpperCAmelCase : List[str] = torch.load(__A , map_location='''cpu''' )['model'] __UpperCAmelCase : Dict = convert_state_dict(__A , __A ) __UpperCAmelCase : Tuple = model.load_state_dict(__A , strict=__A ) assert missing_keys == ["text_model.embeddings.position_ids"] assert (unexpected_keys == ["multi_label_logit_scale"]) or (len(__A ) == 0) # verify result __UpperCAmelCase : List[Any] = CLIPProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) __UpperCAmelCase : Optional[int] = prepare_img() __UpperCAmelCase : Tuple = processor(text=['''a photo of a cat''', '''a photo of a dog'''] , images=__A , padding=__A , return_tensors='''pt''' ) with torch.no_grad(): __UpperCAmelCase : List[str] = model(**__A ) if model_name == "groupvit-gcc-yfcc": __UpperCAmelCase : Any = torch.tensor([[13.35_23, 6.36_29]] ) elif model_name == "groupvit-gcc-redcaps": __UpperCAmelCase : Optional[int] = torch.tensor([[16.18_73, 8.62_30]] ) else: raise ValueError(f'Model name {model_name} not supported.' ) assert torch.allclose(outputs.logits_per_image , __A , atol=1e-3 ) processor.save_pretrained(__A ) model.save_pretrained(__A ) print('''Successfully saved processor and model to''' , __A ) if push_to_hub: print('''Pushing to the hub...''' ) processor.push_to_hub(__A , organization='''nielsr''' ) model.push_to_hub(__A , organization='''nielsr''' ) if __name__ == "__main__": __A =argparse.ArgumentParser() parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to dump the processor and PyTorch model." ) parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to GroupViT checkpoint") parser.add_argument( "--model_name", default="groupvit-gccy-fcc", type=str, help="Name of the model. Expecting either \'groupvit-gcc-yfcc\' or \'groupvit-gcc-redcaps\'", ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model and processor to the 🤗 hub using the provided `model_name`.", ) __A =parser.parse_args() convert_groupvit_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
226
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple=1_3 , SCREAMING_SNAKE_CASE__ : str=7 , SCREAMING_SNAKE_CASE__ : List[str]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : Union[str, Any]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=9_9 , SCREAMING_SNAKE_CASE__ : Optional[Any]=3_2 , SCREAMING_SNAKE_CASE__ : List[str]=5 , SCREAMING_SNAKE_CASE__ : List[Any]=4 , SCREAMING_SNAKE_CASE__ : Tuple=3_7 , SCREAMING_SNAKE_CASE__ : Any="gelu" , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : int=5_1_2 , SCREAMING_SNAKE_CASE__ : int=1_6 , SCREAMING_SNAKE_CASE__ : Optional[int]=2 , SCREAMING_SNAKE_CASE__ : Any=0.02 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE__ : Optional[Any]=4 , SCREAMING_SNAKE_CASE__ : Optional[int]=None , ) -> Any: a_ : Tuple = parent a_ : int = batch_size a_ : Tuple = seq_length a_ : List[Any] = is_training a_ : List[str] = use_token_type_ids a_ : Dict = use_labels a_ : Any = vocab_size a_ : List[str] = hidden_size a_ : Tuple = num_hidden_layers a_ : List[Any] = num_attention_heads a_ : Dict = intermediate_size a_ : Any = hidden_act a_ : List[str] = hidden_dropout_prob a_ : Tuple = attention_probs_dropout_prob a_ : Optional[Any] = max_position_embeddings a_ : List[Any] = type_vocab_size a_ : int = type_sequence_label_size a_ : List[Any] = initializer_range a_ : List[str] = num_labels a_ : Union[str, Any] = num_choices a_ : str = scope a_ : Tuple = self.vocab_size - 1 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Any: a_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a_ : Any = None if self.use_token_type_ids: a_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a_ : List[Any] = None a_ : Union[str, Any] = None a_ : List[Any] = None if self.use_labels: a_ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a_ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) a_ : Union[str, Any] = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) a_ : List[str] = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : List[str] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : List[str] , *SCREAMING_SNAKE_CASE__ : Tuple ) -> Union[str, Any]: a_ : Dict = OpenAIGPTModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : str = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , head_mask=SCREAMING_SNAKE_CASE__ ) a_ : Dict = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ ) a_ : Dict = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Union[str, Any] , *SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Any: a_ : str = OpenAIGPTLMHeadModel(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : Optional[int] = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[Any] , *SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Dict: a_ : int = OpenAIGPTDoubleHeadsModel(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : str = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : Tuple , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , *SCREAMING_SNAKE_CASE__ : str ) -> List[str]: a_ : Any = self.num_labels a_ : Dict = OpenAIGPTForSequenceClassification(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() a_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a_ : Any = model(SCREAMING_SNAKE_CASE__ , token_type_ids=SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: a_ : Optional[Any] = self.prepare_config_and_inputs() ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) : Optional[Any] = config_and_inputs a_ : Optional[int] = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( lowercase__ , lowercase__ , lowercase__ , unittest.TestCase ): snake_case__ : Tuple = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) snake_case__ : List[str] = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly snake_case__ : Dict = ( { '''feature-extraction''': OpenAIGPTModel, '''text-classification''': OpenAIGPTForSequenceClassification, '''text-generation''': OpenAIGPTLMHeadModel, '''zero-shot''': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE ( self : List[str] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[str] ) -> Dict: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def SCREAMING_SNAKE_CASE ( self : int , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Any=False ) -> List[str]: a_ : str = super()._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": a_ : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=SCREAMING_SNAKE_CASE__ , ) a_ : str = inputs_dict['labels'] a_ : Optional[int] = inputs_dict['labels'] a_ : Optional[int] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=SCREAMING_SNAKE_CASE__ , ) a_ : Union[str, Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=SCREAMING_SNAKE_CASE__ ) return inputs_dict def SCREAMING_SNAKE_CASE ( self : str ) -> List[Any]: a_ : str = OpenAIGPTModelTester(self ) a_ : int = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , n_embd=3_7 ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Tuple: a_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: a_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Any ) -> Optional[Any]: a_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[Any]: a_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ : str = OpenAIGPTModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: a_ : Dict = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt' ) model.to(SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=SCREAMING_SNAKE_CASE__ ) # the president is a_ : Tuple = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the a_ : Dict = model.generate(SCREAMING_SNAKE_CASE__ , do_sample=SCREAMING_SNAKE_CASE__ ) self.assertListEqual(output_ids[0].tolist() , SCREAMING_SNAKE_CASE__ )
32
0
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed A__ = { 'distilbert': (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), 'roberta': (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), 'bert': (BertConfig, BertForMaskedLM, BertTokenizer), 'gpt2': (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def _lowerCAmelCase ( __lowerCAmelCase ) -> int: """simple docstring""" assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> Optional[Any]: """simple docstring""" if args.student_type == "roberta": snake_case__ : List[Any] = False elif args.student_type == "gpt2": snake_case__ : str = False def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> Dict: """simple docstring""" if args.student_type == "roberta": snake_case__ : Union[str, Any] = False def _lowerCAmelCase ( ) -> List[str]: """simple docstring""" snake_case__ : List[str] = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=__A , required=__A , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=__A , required=__A , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=__A , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=__A , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=__A , required=__A , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=__A , type=__A , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=__A , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=__A , required=__A , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=__A , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=__A , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=__A , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=__A , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=__A , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=__A , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.15 , type=__A , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=__A , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=__A , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=__A , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=__A , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=__A , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=__A , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=__A , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=__A , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.05 , type=__A , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=__A , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5E-4 , type=__A , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1E-6 , type=__A , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=__A , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.02 , type=__A , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=__A , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=__A , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=__A , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=__A , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=__A , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=__A , default=4000 , help='''Checkpoint interval.''' ) snake_case__ : Any = parser.parse_args() sanity_checks(__A ) # ARGS # init_gpu_params(__A ) set_seed(__A ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(f"""Param: {args}""" ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(__A ) , __A , indent=4 ) git_log(args.dump_path ) snake_case__ : str = MODEL_CLASSES[args.student_type] snake_case__ : List[Any] = MODEL_CLASSES[args.teacher_type] # TOKENIZER # snake_case__ : Tuple = teacher_tokenizer_class.from_pretrained(args.teacher_name ) snake_case__ : List[str] = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): snake_case__ : Dict = tokenizer.all_special_tokens.index(__A ) snake_case__ : str = tokenizer.all_special_ids[idx] logger.info(f"""Special tokens {special_tok_ids}""" ) snake_case__ : List[str] = special_tok_ids snake_case__ : Optional[int] = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"""Loading data from {args.data_file}""" ) with open(args.data_file , '''rb''' ) as fp: snake_case__ : int = pickle.load(__A ) if args.mlm: logger.info(f"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts , '''rb''' ) as fp: snake_case__ : Dict = pickle.load(__A ) snake_case__ : str = np.maximum(__A , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): snake_case__ : Tuple = 0.0 # do not predict special tokens snake_case__ : str = torch.from_numpy(__A ) else: snake_case__ : List[str] = None snake_case__ : List[Any] = LmSeqsDataset(params=__A , data=__A ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f"""Loading student config from {args.student_config}""" ) snake_case__ : int = student_config_class.from_pretrained(args.student_config ) snake_case__ : Any = True if args.student_pretrained_weights is not None: logger.info(f"""Loading pretrained weights from {args.student_pretrained_weights}""" ) snake_case__ : int = student_model_class.from_pretrained(args.student_pretrained_weights , config=__A ) else: snake_case__ : Dict = student_model_class(__A ) if args.n_gpu > 0: student.to(f"""cuda:{args.local_rank}""" ) logger.info('''Student loaded.''' ) # TEACHER # snake_case__ : int = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=__A ) if args.n_gpu > 0: teacher.to(f"""cuda:{args.local_rank}""" ) logger.info(f"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__A , __A ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__A , __A ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() snake_case__ : List[str] = Distiller( params=__A , dataset=__A , token_probs=__A , student=__A , teacher=__A ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
230
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ : Optional[int] = { 'facebook/mask2former-swin-small-coco-instance': ( 'https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json' ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } UpperCAmelCase_ : List[str] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Any = '''mask2former''' snake_case__ : Any = ['''swin'''] snake_case__ : str = {'''hidden_size''': '''hidden_dim'''} def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Optional[Dict] = None , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 2_5_6 , SCREAMING_SNAKE_CASE__ : int = 1_0_2_4 , SCREAMING_SNAKE_CASE__ : str = "relu" , SCREAMING_SNAKE_CASE__ : int = 6 , SCREAMING_SNAKE_CASE__ : int = 1_0 , SCREAMING_SNAKE_CASE__ : int = 8 , SCREAMING_SNAKE_CASE__ : float = 0.0 , SCREAMING_SNAKE_CASE__ : int = 2_0_4_8 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : int = 4 , SCREAMING_SNAKE_CASE__ : int = 2_5_5 , SCREAMING_SNAKE_CASE__ : int = 1_0_0 , SCREAMING_SNAKE_CASE__ : float = 0.1 , SCREAMING_SNAKE_CASE__ : float = 2.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : float = 5.0 , SCREAMING_SNAKE_CASE__ : int = 1_2_5_4_4 , SCREAMING_SNAKE_CASE__ : float = 3.0 , SCREAMING_SNAKE_CASE__ : float = 0.75 , SCREAMING_SNAKE_CASE__ : float = 0.02 , SCREAMING_SNAKE_CASE__ : float = 1.0 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : List[int] = [4, 8, 1_6, 3_2] , SCREAMING_SNAKE_CASE__ : bool = None , **SCREAMING_SNAKE_CASE__ : int , ) -> List[Any]: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) a_ : Dict = CONFIG_MAPPING['swin']( image_size=2_2_4 , in_channels=3 , patch_size=4 , embed_dim=9_6 , depths=[2, 2, 1_8, 2] , num_heads=[3, 6, 1_2, 2_4] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=SCREAMING_SNAKE_CASE__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): a_ : Any = backbone_config.pop('model_type' ) a_ : Optional[Any] = CONFIG_MAPPING[backbone_model_type] a_ : List[str] = config_class.from_dict(SCREAMING_SNAKE_CASE__ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. """ F"""Supported model types: {",".join(self.backbones_supported )}""" ) a_ : Dict = backbone_config a_ : List[str] = feature_size a_ : List[str] = mask_feature_size a_ : int = hidden_dim a_ : Dict = encoder_feedforward_dim a_ : str = activation_function a_ : List[str] = encoder_layers a_ : List[str] = decoder_layers a_ : Dict = num_attention_heads a_ : str = dropout a_ : Tuple = dim_feedforward a_ : List[str] = pre_norm a_ : Optional[int] = enforce_input_projection a_ : Any = common_stride a_ : Optional[int] = ignore_value a_ : int = num_queries a_ : Tuple = no_object_weight a_ : Dict = class_weight a_ : Optional[int] = mask_weight a_ : Optional[int] = dice_weight a_ : str = train_num_points a_ : List[str] = oversample_ratio a_ : List[Any] = importance_sample_ratio a_ : Any = init_std a_ : Union[str, Any] = init_xavier_std a_ : Union[str, Any] = use_auxiliary_loss a_ : Dict = feature_strides a_ : List[str] = output_auxiliary_logits a_ : Dict = decoder_layers super().__init__(**SCREAMING_SNAKE_CASE__ ) @classmethod def SCREAMING_SNAKE_CASE ( cls : str , SCREAMING_SNAKE_CASE__ : PretrainedConfig , **SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> List[Any]: return cls( backbone_config=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict[str, any]: a_ : Optional[int] = copy.deepcopy(self.__dict__ ) a_ : List[Any] = self.backbone_config.to_dict() a_ : Optional[Any] = self.__class__.model_type return output
32
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase : str = logging.get_logger(__name__) _lowercase : List[str] = { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json', # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _UpperCAmelCase ( lowercase__ ): a__ : Optional[int] = '''convbert''' def __init__( self : List[Any] , _lowercase : Optional[int]=3_05_22 , _lowercase : Dict=7_68 , _lowercase : Optional[int]=12 , _lowercase : Union[str, Any]=12 , _lowercase : str=30_72 , _lowercase : Dict="gelu" , _lowercase : Dict=0.1 , _lowercase : Tuple=0.1 , _lowercase : List[str]=5_12 , _lowercase : Optional[Any]=2 , _lowercase : List[Any]=0.02 , _lowercase : Any=1E-12 , _lowercase : int=1 , _lowercase : int=0 , _lowercase : Optional[int]=2 , _lowercase : Optional[int]=7_68 , _lowercase : Union[str, Any]=2 , _lowercase : List[Any]=9 , _lowercase : List[Any]=1 , _lowercase : Dict=None , **_lowercase : List[Any] , ): super().__init__( pad_token_id=SCREAMING_SNAKE_CASE__ , bos_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) __UpperCAmelCase = vocab_size __UpperCAmelCase = hidden_size __UpperCAmelCase = num_hidden_layers __UpperCAmelCase = num_attention_heads __UpperCAmelCase = intermediate_size __UpperCAmelCase = hidden_act __UpperCAmelCase = hidden_dropout_prob __UpperCAmelCase = attention_probs_dropout_prob __UpperCAmelCase = max_position_embeddings __UpperCAmelCase = type_vocab_size __UpperCAmelCase = initializer_range __UpperCAmelCase = layer_norm_eps __UpperCAmelCase = embedding_size __UpperCAmelCase = head_ratio __UpperCAmelCase = conv_kernel_size __UpperCAmelCase = num_groups __UpperCAmelCase = classifier_dropout class _UpperCAmelCase ( lowercase__ ): @property def a ( self : List[str] ): if self.task == "multiple-choice": __UpperCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __UpperCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
332
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : Union[str, Any] = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : List[str] = '''switch_transformers''' snake_case__ : Optional[int] = ['''past_key_values'''] snake_case__ : Optional[Any] = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self : Dict , SCREAMING_SNAKE_CASE__ : Optional[int]=3_2_1_2_8 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=7_6_8 , SCREAMING_SNAKE_CASE__ : Optional[Any]=6_4 , SCREAMING_SNAKE_CASE__ : List[str]=2_0_4_8 , SCREAMING_SNAKE_CASE__ : Dict=6_4 , SCREAMING_SNAKE_CASE__ : List[Any]=1_2 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=1_2 , SCREAMING_SNAKE_CASE__ : str=3 , SCREAMING_SNAKE_CASE__ : Tuple=1_2 , SCREAMING_SNAKE_CASE__ : Tuple=8 , SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.01 , SCREAMING_SNAKE_CASE__ : str="float32" , SCREAMING_SNAKE_CASE__ : str=False , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3_2 , SCREAMING_SNAKE_CASE__ : Dict=1_2_8 , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : Dict=1E-6 , SCREAMING_SNAKE_CASE__ : Dict=0.001 , SCREAMING_SNAKE_CASE__ : Any=0.001 , SCREAMING_SNAKE_CASE__ : Optional[int]=1.0 , SCREAMING_SNAKE_CASE__ : Any="relu" , SCREAMING_SNAKE_CASE__ : List[Any]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : Optional[int]=1 , **SCREAMING_SNAKE_CASE__ : Dict , ) -> Optional[Any]: a_ : Optional[int] = vocab_size a_ : List[str] = d_model a_ : Tuple = d_kv a_ : Optional[Any] = d_ff a_ : List[Any] = num_sparse_encoder_layers a_ : Any = num_layers a_ : str = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry a_ : List[Any] = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: a_ : Optional[int] = self.num_layers // self.num_sparse_encoder_layers else: a_ : List[Any] = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: a_ : Union[str, Any] = self.num_decoder_layers // self.num_sparse_decoder_layers else: a_ : List[str] = self.num_decoder_layers # HACK: this will create 0 sparse layers a_ : Dict = num_heads a_ : str = num_experts a_ : Any = expert_capacity a_ : List[Any] = router_bias a_ : str = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) a_ : Optional[int] = router_dtype a_ : int = router_ignore_padding_tokens a_ : Any = relative_attention_num_buckets a_ : List[str] = relative_attention_max_distance a_ : Optional[Any] = dropout_rate a_ : Tuple = layer_norm_epsilon a_ : Dict = initializer_factor a_ : Any = feed_forward_proj a_ : Tuple = use_cache a_ : str = add_router_probs a_ : Optional[int] = router_z_loss_coef a_ : List[str] = router_aux_loss_coef a_ : int = self.feed_forward_proj.split('-' ) a_ : int = act_info[-1] a_ : Optional[int] = act_info[0] == 'gated' if len(SCREAMING_SNAKE_CASE__ ) > 1 and act_info[0] != "gated" or len(SCREAMING_SNAKE_CASE__ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": a_ : Any = 'gelu_new' super().__init__( pad_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , is_encoder_decoder=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , )
32
0
__A : Optional[int] = { 'Pillow': 'Pillow<10.0.0', 'accelerate': 'accelerate>=0.20.3', 'av': 'av==9.2.0', 'beautifulsoup4': 'beautifulsoup4', 'black': 'black~=23.1', 'codecarbon': 'codecarbon==1.2.0', 'cookiecutter': 'cookiecutter==1.7.3', 'dataclasses': 'dataclasses', 'datasets': 'datasets!=2.5.0', 'decord': 'decord==0.6.0', 'deepspeed': 'deepspeed>=0.9.3', 'diffusers': 'diffusers', 'dill': 'dill<0.3.5', 'evaluate': 'evaluate>=0.2.0', 'fairscale': 'fairscale>0.3', 'faiss-cpu': 'faiss-cpu', 'fastapi': 'fastapi', 'filelock': 'filelock', 'flax': 'flax>=0.4.1,<=0.7.0', 'ftfy': 'ftfy', 'fugashi': 'fugashi>=1.0', 'GitPython': 'GitPython<3.1.19', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.14.1,<1.0', 'importlib_metadata': 'importlib_metadata', 'ipadic': 'ipadic>=1.0.0,<2.0', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2,<=0.4.13', 'jaxlib': 'jaxlib>=0.1.65,<=0.4.13', 'jieba': 'jieba', 'kenlm': 'kenlm', 'keras-nlp': 'keras-nlp>=0.3.1', 'librosa': 'librosa', 'nltk': 'nltk', 'natten': 'natten>=0.14.6', 'numpy': 'numpy>=1.17', 'onnxconverter-common': 'onnxconverter-common', 'onnxruntime-tools': 'onnxruntime-tools>=1.4.2', 'onnxruntime': 'onnxruntime>=1.4.0', 'opencv-python': 'opencv-python', 'optuna': 'optuna', 'optax': 'optax>=0.0.8,<=0.1.4', 'packaging': 'packaging>=20.0', 'parameterized': 'parameterized', 'phonemizer': 'phonemizer', 'protobuf': 'protobuf', 'psutil': 'psutil', 'pyyaml': 'pyyaml>=5.1', 'pydantic': 'pydantic<2', 'pytest': 'pytest>=7.2.0', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'python': 'python>=3.8.0', 'ray[tune]': 'ray[tune]', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'rhoknp': 'rhoknp>=1.1.0,<1.3.1', 'rjieba': 'rjieba', 'rouge-score': 'rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1', 'ruff': 'ruff>=0.0.241,<=0.0.259', 'sacrebleu': 'sacrebleu>=1.4.12,<2.0.0', 'sacremoses': 'sacremoses', 'safetensors': 'safetensors>=0.3.1', 'sagemaker': 'sagemaker>=2.31.0', 'scikit-learn': 'scikit-learn', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'sigopt': 'sigopt', 'starlette': 'starlette', 'sudachipy': 'sudachipy>=0.6.6', 'sudachidict_core': 'sudachidict_core>=20220729', 'tensorflow-cpu': 'tensorflow-cpu>=2.6,<2.14', 'tensorflow': 'tensorflow>=2.6,<2.14', 'tensorflow-text': 'tensorflow-text<2.14', 'tf2onnx': 'tf2onnx', 'timeout-decorator': 'timeout-decorator', 'timm': 'timm', 'tokenizers': 'tokenizers>=0.11.1,!=0.11.3,<0.14', 'torch': 'torch>=1.9,!=1.12.0', 'torchaudio': 'torchaudio', 'torchvision': 'torchvision', 'pyctcdecode': 'pyctcdecode>=0.4.0', 'tqdm': 'tqdm>=4.27', 'unidic': 'unidic>=1.0.2', 'unidic_lite': 'unidic_lite>=1.0.7', 'urllib3': 'urllib3<2.0.0', 'uvicorn': 'uvicorn', }
154
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool UpperCAmelCase_ : Tuple = { 'Acehnese Arabic': 'ace_Arab', 'Acehnese Latin': 'ace_Latn', 'Mesopotamian Arabic': 'acm_Arab', 'Ta\'izzi-Adeni Arabic': 'acq_Arab', 'Tunisian Arabic': 'aeb_Arab', 'Afrikaans': 'afr_Latn', 'South Levantine Arabic': 'ajp_Arab', 'Akan': 'aka_Latn', 'Amharic': 'amh_Ethi', 'North Levantine Arabic': 'apc_Arab', 'Modern Standard Arabic': 'arb_Arab', 'Modern Standard Arabic Romanized': 'arb_Latn', 'Najdi Arabic': 'ars_Arab', 'Moroccan Arabic': 'ary_Arab', 'Egyptian Arabic': 'arz_Arab', 'Assamese': 'asm_Beng', 'Asturian': 'ast_Latn', 'Awadhi': 'awa_Deva', 'Central Aymara': 'ayr_Latn', 'South Azerbaijani': 'azb_Arab', 'North Azerbaijani': 'azj_Latn', 'Bashkir': 'bak_Cyrl', 'Bambara': 'bam_Latn', 'Balinese': 'ban_Latn', 'Belarusian': 'bel_Cyrl', 'Bemba': 'bem_Latn', 'Bengali': 'ben_Beng', 'Bhojpuri': 'bho_Deva', 'Banjar Arabic': 'bjn_Arab', 'Banjar Latin': 'bjn_Latn', 'Standard Tibetan': 'bod_Tibt', 'Bosnian': 'bos_Latn', 'Buginese': 'bug_Latn', 'Bulgarian': 'bul_Cyrl', 'Catalan': 'cat_Latn', 'Cebuano': 'ceb_Latn', 'Czech': 'ces_Latn', 'Chokwe': 'cjk_Latn', 'Central Kurdish': 'ckb_Arab', 'Crimean Tatar': 'crh_Latn', 'Welsh': 'cym_Latn', 'Danish': 'dan_Latn', 'German': 'deu_Latn', 'Southwestern Dinka': 'dik_Latn', 'Dyula': 'dyu_Latn', 'Dzongkha': 'dzo_Tibt', 'Greek': 'ell_Grek', 'English': 'eng_Latn', 'Esperanto': 'epo_Latn', 'Estonian': 'est_Latn', 'Basque': 'eus_Latn', 'Ewe': 'ewe_Latn', 'Faroese': 'fao_Latn', 'Fijian': 'fij_Latn', 'Finnish': 'fin_Latn', 'Fon': 'fon_Latn', 'French': 'fra_Latn', 'Friulian': 'fur_Latn', 'Nigerian Fulfulde': 'fuv_Latn', 'Scottish Gaelic': 'gla_Latn', 'Irish': 'gle_Latn', 'Galician': 'glg_Latn', 'Guarani': 'grn_Latn', 'Gujarati': 'guj_Gujr', 'Haitian Creole': 'hat_Latn', 'Hausa': 'hau_Latn', 'Hebrew': 'heb_Hebr', 'Hindi': 'hin_Deva', 'Chhattisgarhi': 'hne_Deva', 'Croatian': 'hrv_Latn', 'Hungarian': 'hun_Latn', 'Armenian': 'hye_Armn', 'Igbo': 'ibo_Latn', 'Ilocano': 'ilo_Latn', 'Indonesian': 'ind_Latn', 'Icelandic': 'isl_Latn', 'Italian': 'ita_Latn', 'Javanese': 'jav_Latn', 'Japanese': 'jpn_Jpan', 'Kabyle': 'kab_Latn', 'Jingpho': 'kac_Latn', 'Kamba': 'kam_Latn', 'Kannada': 'kan_Knda', 'Kashmiri Arabic': 'kas_Arab', 'Kashmiri Devanagari': 'kas_Deva', 'Georgian': 'kat_Geor', 'Central Kanuri Arabic': 'knc_Arab', 'Central Kanuri Latin': 'knc_Latn', 'Kazakh': 'kaz_Cyrl', 'Kabiyè': 'kbp_Latn', 'Kabuverdianu': 'kea_Latn', 'Khmer': 'khm_Khmr', 'Kikuyu': 'kik_Latn', 'Kinyarwanda': 'kin_Latn', 'Kyrgyz': 'kir_Cyrl', 'Kimbundu': 'kmb_Latn', 'Northern Kurdish': 'kmr_Latn', 'Kikongo': 'kon_Latn', 'Korean': 'kor_Hang', 'Lao': 'lao_Laoo', 'Ligurian': 'lij_Latn', 'Limburgish': 'lim_Latn', 'Lingala': 'lin_Latn', 'Lithuanian': 'lit_Latn', 'Lombard': 'lmo_Latn', 'Latgalian': 'ltg_Latn', 'Luxembourgish': 'ltz_Latn', 'Luba-Kasai': 'lua_Latn', 'Ganda': 'lug_Latn', 'Luo': 'luo_Latn', 'Mizo': 'lus_Latn', 'Standard Latvian': 'lvs_Latn', 'Magahi': 'mag_Deva', 'Maithili': 'mai_Deva', 'Malayalam': 'mal_Mlym', 'Marathi': 'mar_Deva', 'Minangkabau Arabic ': 'min_Arab', 'Minangkabau Latin': 'min_Latn', 'Macedonian': 'mkd_Cyrl', 'Plateau Malagasy': 'plt_Latn', 'Maltese': 'mlt_Latn', 'Meitei Bengali': 'mni_Beng', 'Halh Mongolian': 'khk_Cyrl', 'Mossi': 'mos_Latn', 'Maori': 'mri_Latn', 'Burmese': 'mya_Mymr', 'Dutch': 'nld_Latn', 'Norwegian Nynorsk': 'nno_Latn', 'Norwegian Bokmål': 'nob_Latn', 'Nepali': 'npi_Deva', 'Northern Sotho': 'nso_Latn', 'Nuer': 'nus_Latn', 'Nyanja': 'nya_Latn', 'Occitan': 'oci_Latn', 'West Central Oromo': 'gaz_Latn', 'Odia': 'ory_Orya', 'Pangasinan': 'pag_Latn', 'Eastern Panjabi': 'pan_Guru', 'Papiamento': 'pap_Latn', 'Western Persian': 'pes_Arab', 'Polish': 'pol_Latn', 'Portuguese': 'por_Latn', 'Dari': 'prs_Arab', 'Southern Pashto': 'pbt_Arab', 'Ayacucho Quechua': 'quy_Latn', 'Romanian': 'ron_Latn', 'Rundi': 'run_Latn', 'Russian': 'rus_Cyrl', 'Sango': 'sag_Latn', 'Sanskrit': 'san_Deva', 'Santali': 'sat_Olck', 'Sicilian': 'scn_Latn', 'Shan': 'shn_Mymr', 'Sinhala': 'sin_Sinh', 'Slovak': 'slk_Latn', 'Slovenian': 'slv_Latn', 'Samoan': 'smo_Latn', 'Shona': 'sna_Latn', 'Sindhi': 'snd_Arab', 'Somali': 'som_Latn', 'Southern Sotho': 'sot_Latn', 'Spanish': 'spa_Latn', 'Tosk Albanian': 'als_Latn', 'Sardinian': 'srd_Latn', 'Serbian': 'srp_Cyrl', 'Swati': 'ssw_Latn', 'Sundanese': 'sun_Latn', 'Swedish': 'swe_Latn', 'Swahili': 'swh_Latn', 'Silesian': 'szl_Latn', 'Tamil': 'tam_Taml', 'Tatar': 'tat_Cyrl', 'Telugu': 'tel_Telu', 'Tajik': 'tgk_Cyrl', 'Tagalog': 'tgl_Latn', 'Thai': 'tha_Thai', 'Tigrinya': 'tir_Ethi', 'Tamasheq Latin': 'taq_Latn', 'Tamasheq Tifinagh': 'taq_Tfng', 'Tok Pisin': 'tpi_Latn', 'Tswana': 'tsn_Latn', 'Tsonga': 'tso_Latn', 'Turkmen': 'tuk_Latn', 'Tumbuka': 'tum_Latn', 'Turkish': 'tur_Latn', 'Twi': 'twi_Latn', 'Central Atlas Tamazight': 'tzm_Tfng', 'Uyghur': 'uig_Arab', 'Ukrainian': 'ukr_Cyrl', 'Umbundu': 'umb_Latn', 'Urdu': 'urd_Arab', 'Northern Uzbek': 'uzn_Latn', 'Venetian': 'vec_Latn', 'Vietnamese': 'vie_Latn', 'Waray': 'war_Latn', 'Wolof': 'wol_Latn', 'Xhosa': 'xho_Latn', 'Eastern Yiddish': 'ydd_Hebr', 'Yoruba': 'yor_Latn', 'Yue Chinese': 'yue_Hant', 'Chinese Simplified': 'zho_Hans', 'Chinese Traditional': 'zho_Hant', 'Standard Malay': 'zsm_Latn', 'Zulu': 'zul_Latn', } class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : str = '''facebook/nllb-200-distilled-600M''' snake_case__ : Union[str, Any] = ( '''This is a tool that translates text from a language to another. It takes three inputs: `text`, which should ''' '''be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, ''' '''which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in ''' '''plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.''' ) snake_case__ : Optional[Any] = '''translator''' snake_case__ : Tuple = AutoTokenizer snake_case__ : Union[str, Any] = AutoModelForSeqaSeqLM snake_case__ : Dict = LANGUAGE_CODES snake_case__ : str = ['''text''', '''text''', '''text'''] snake_case__ : Tuple = ['''text'''] def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Tuple: if src_lang not in self.lang_to_code: raise ValueError(F"""{src_lang} is not a supported language.""" ) if tgt_lang not in self.lang_to_code: raise ValueError(F"""{tgt_lang} is not a supported language.""" ) a_ : str = self.lang_to_code[src_lang] a_ : Any = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( SCREAMING_SNAKE_CASE__ , return_tensors='pt' , src_lang=SCREAMING_SNAKE_CASE__ , tgt_lang=SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : int , SCREAMING_SNAKE_CASE__ : Tuple ) -> Any: return self.model.generate(**SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[str] ) -> Dict: return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=SCREAMING_SNAKE_CASE__ )
32
0
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __a: Tuple = logging.get_logger(__name__) class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase=None , __lowerCAmelCase=None ) -> List[Any]: if not conversation_id: lowercase__ : Any = uuid.uuida() if past_user_inputs is None: lowercase__ : int = [] if generated_responses is None: lowercase__ : int = [] lowercase__ : uuid.UUID = conversation_id lowercase__ : List[str] = past_user_inputs lowercase__ : List[str] = generated_responses lowercase__ : Optional[str] = text def __eq__( self , __lowerCAmelCase ) -> Tuple: if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase = False ) -> Optional[Any]: if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) lowercase__ : Optional[int] = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: lowercase__ : Optional[Any] = text def _lowerCAmelCase( self ) -> List[str]: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) lowercase__ : List[Any] = None def _lowerCAmelCase( self , __lowerCAmelCase ) -> Dict: self.generated_responses.append(SCREAMING_SNAKE_CASE__ ) def _lowerCAmelCase( self ) -> Optional[Any]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ) -> Dict: lowercase__ : str = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): lowercase__ : List[str] = 'user' if is_user else 'bot' output += F"""{name} >> {text} \n""" return output @add_end_docstrings( lowercase__ , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class UpperCAmelCase ( lowercase__ ): '''simple docstring''' def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ) -> Optional[Any]: super().__init__(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if self.tokenizer.pad_token_id is None: lowercase__ : Dict = self.tokenizer.eos_token def _lowerCAmelCase( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase ) -> Union[str, Any]: lowercase__ : Optional[int] = {} lowercase__ : Optional[Any] = {} lowercase__ : Dict = {} if min_length_for_response is not None: lowercase__ : Optional[int] = min_length_for_response if minimum_tokens is not None: lowercase__ : Union[str, Any] = minimum_tokens if "max_length" in generate_kwargs: lowercase__ : Any = generate_kwargs['max_length'] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: lowercase__ : List[Any] = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(SCREAMING_SNAKE_CASE__ ) return preprocess_params, forward_params, postprocess_params def __call__( self , __lowerCAmelCase , __lowerCAmelCase=0 , **__lowerCAmelCase ) -> List[str]: lowercase__ : Dict = super().__call__(SCREAMING_SNAKE_CASE__ , num_workers=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and len(SCREAMING_SNAKE_CASE__ ) == 1: return outputs[0] return outputs def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError('''ConversationalPipeline, expects Conversation as inputs''' ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ '''Add user inputs with the conversation\'s `add_user_input` method''' ) if hasattr(self.tokenizer , '''_build_conversation_input_ids''' ): lowercase__ : str = self.tokenizer._build_conversation_input_ids(SCREAMING_SNAKE_CASE__ ) else: # If the tokenizer cannot handle conversations, we default to only the old version lowercase__ : Any = self._legacy_parse_and_tokenize(SCREAMING_SNAKE_CASE__ ) if self.framework == "pt": lowercase__ : Any = torch.LongTensor([input_ids] ) elif self.framework == "tf": lowercase__ : Dict = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=10 , **__lowerCAmelCase ) -> Any: lowercase__ : Union[str, Any] = generate_kwargs.get('''max_length''' , self.model.config.max_length ) lowercase__ : int = model_inputs['input_ids'].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) lowercase__ : Any = max_length - minimum_tokens lowercase__ : Tuple = model_inputs['input_ids'][:, -trim:] if "attention_mask" in model_inputs: lowercase__ : Tuple = model_inputs['attention_mask'][:, -trim:] lowercase__ : Dict = model_inputs.pop('''conversation''' ) lowercase__ : List[Any] = max_length lowercase__ : Optional[int] = self.model.generate(**SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if self.model.config.is_encoder_decoder: lowercase__ : Optional[int] = 1 else: lowercase__ : List[str] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=True ) -> Optional[Any]: lowercase__ : Optional[int] = model_outputs['output_ids'] lowercase__ : List[Any] = self.tokenizer.decode( output_ids[0] , skip_special_tokens=SCREAMING_SNAKE_CASE__ , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE__ , ) lowercase__ : Optional[int] = model_outputs['conversation'] conversation.mark_processed() conversation.append_response(SCREAMING_SNAKE_CASE__ ) return conversation def _lowerCAmelCase( self , __lowerCAmelCase ) -> Dict: lowercase__ : Optional[int] = self.tokenizer.eos_token_id lowercase__ : List[Any] = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ ) ) if len(SCREAMING_SNAKE_CASE__ ) > self.tokenizer.model_max_length: lowercase__ : str = input_ids[-self.tokenizer.model_max_length :] return input_ids
198
UpperCAmelCase_ : Optional[int] = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCAmelCase_ : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] UpperCAmelCase_ : str = { 0: 'Sunday', 1: 'Monday', 2: 'Tuesday', 3: 'Wednesday', 4: 'Thursday', 5: 'Friday', 6: 'Saturday', } def SCREAMING_SNAKE_CASE_ ( __A : int , __A : int , __A : int ) -> str: """simple docstring""" assert len(str(__A ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: a_ : List[str] = year // 1_00 a_ : Optional[int] = (5 * (century % 4) + 2) % 7 a_ : List[str] = year % 1_00 a_ : str = centurian % 12 a_ : List[str] = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 a_ : Any = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) a_ : Any = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
32
0
"""simple docstring""" from __future__ import annotations import requests A = set( '''approved_at_utc approved_by author_flair_background_color\nauthor_flair_css_class author_flair_richtext author_flair_template_id author_fullname\nauthor_premium can_mod_post category clicked content_categories created_utc downs\nedited gilded gildings hidden hide_score is_created_from_ads_ui is_meta\nis_original_content is_reddit_media_domain is_video link_flair_css_class\nlink_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title\nname permalink pwls quarantine saved score secure_media secure_media_embed selftext\nsubreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type\ntotal_awards_received ups upvote_ratio url user_reports'''.split() ) def __A ( a_ :str , a_ :int = 1 , a_ :str = "new" , a_ :list | None = None) -> dict: __a : List[Any] = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(__A) - valid_terms)): __a : List[Any] = F"""Invalid search term: {invalid_search_terms}""" raise ValueError(__A) __a : Dict = requests.get( F"""https://reddit.com/r/{subreddit}/{age}.json?limit={limit}""" , headers={'''User-agent''': '''A random string'''} , ) if response.status_code == 4_29: raise requests.HTTPError __a : Dict = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(__A)} __a : List[Any] = {} for id_ in range(__A): __a : Any = { item: data['data']['children'][id_]['data'][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
160
import math import flax.linen as nn import jax.numpy as jnp def SCREAMING_SNAKE_CASE_ ( __A : jnp.ndarray , __A : int , __A : float = 1 , __A : float = 1 , __A : float = 1.0e4 , __A : bool = False , __A : float = 1.0 , ) -> jnp.ndarray: """simple docstring""" assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, F"""Embedding dimension {embedding_dim} should be even""" a_ : int = float(embedding_dim // 2 ) a_ : str = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) a_ : Optional[int] = min_timescale * jnp.exp(jnp.arange(__A , dtype=jnp.floataa ) * -log_timescale_increment ) a_ : Optional[int] = jnp.expand_dims(__A , 1 ) * jnp.expand_dims(__A , 0 ) # scale embeddings a_ : str = scale * emb if flip_sin_to_cos: a_ : str = jnp.concatenate([jnp.cos(__A ), jnp.sin(__A )] , axis=1 ) else: a_ : Any = jnp.concatenate([jnp.sin(__A ), jnp.cos(__A )] , axis=1 ) a_ : Optional[int] = jnp.reshape(__A , [jnp.shape(__A )[0], embedding_dim] ) return signal class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int = 32 snake_case__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self : Tuple , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: a_ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(SCREAMING_SNAKE_CASE__ ) a_ : Tuple = nn.silu(SCREAMING_SNAKE_CASE__ ) a_ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(SCREAMING_SNAKE_CASE__ ) return temb class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int = 32 snake_case__ : bool = False snake_case__ : float = 1 @nn.compact def __call__( self : str , SCREAMING_SNAKE_CASE__ : int ) -> Tuple: return get_sinusoidal_embeddings( SCREAMING_SNAKE_CASE__ , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
32
0
from __future__ import annotations import math import random from typing import Any class __A: def __init__( self ) -> None: '''simple docstring''' __a = [] __a = 0 __a = 0 def SCREAMING_SNAKE_CASE_ ( self ) -> bool: '''simple docstring''' return self.head == self.tail def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' self.data.append(SCREAMING_SNAKE_CASE__ ) __a = self.tail + 1 def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' __a = self.data[self.head] __a = self.head + 1 return ret def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return self.tail - self.head def SCREAMING_SNAKE_CASE_ ( self ) -> None: '''simple docstring''' print(self.data ) print('''**************''' ) print(self.data[self.head : self.tail] ) class __A: def __init__( self , _snake_case ) -> None: '''simple docstring''' __a = data __a = None __a = None __a = 1 def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' return self.data def SCREAMING_SNAKE_CASE_ ( self ) -> MyNode | None: '''simple docstring''' return self.left def SCREAMING_SNAKE_CASE_ ( self ) -> MyNode | None: '''simple docstring''' return self.right def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return self.height def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' __a = data def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' __a = node def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' __a = node def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' __a = height def __lowerCAmelCase ( a__ ) -> int: if node is None: return 0 return node.get_height() def __lowerCAmelCase ( a__ , a__ ) -> int: if a > b: return a return b def __lowerCAmelCase ( a__ ) -> MyNode: print('''left rotation node:''' , node.get_data() ) __a = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(__A ) __a = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__A ) __a = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(__A ) return ret def __lowerCAmelCase ( a__ ) -> MyNode: print('''right rotation node:''' , node.get_data() ) __a = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(__A ) __a = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__A ) __a = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(__A ) return ret def __lowerCAmelCase ( a__ ) -> MyNode: __a = node.get_left() assert left_child is not None node.set_left(left_rotation(__A ) ) return right_rotation(__A ) def __lowerCAmelCase ( a__ ) -> MyNode: __a = node.get_right() assert right_child is not None node.set_right(right_rotation(__A ) ) return left_rotation(__A ) def __lowerCAmelCase ( a__ , a__ ) -> MyNode | None: if node is None: return MyNode(__A ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , __A ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected __a = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child __a = right_rotation(__A ) else: __a = lr_rotation(__A ) else: node.set_right(insert_node(node.get_right() , __A ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: __a = node.get_right() assert right_child is not None if data < right_child.get_data(): __a = rl_rotation(__A ) else: __a = left_rotation(__A ) __a = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__A ) return node def __lowerCAmelCase ( a__ ) -> Any: while True: __a = root.get_right() if right_child is None: break __a = right_child return root.get_data() def __lowerCAmelCase ( a__ ) -> Any: while True: __a = root.get_left() if left_child is None: break __a = left_child return root.get_data() def __lowerCAmelCase ( a__ , a__ ) -> MyNode | None: __a = root.get_left() __a = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: __a = get_left_most(__A ) root.set_data(__A ) root.set_right(del_node(__A , __A ) ) elif left_child is not None: __a = left_child elif right_child is not None: __a = right_child else: return None elif root.get_data() > data: if left_child is None: print('''No such data''' ) return root else: root.set_left(del_node(__A , __A ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(__A , __A ) ) if get_height(__A ) - get_height(__A ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): __a = left_rotation(__A ) else: __a = rl_rotation(__A ) elif get_height(__A ) - get_height(__A ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): __a = right_rotation(__A ) else: __a = lr_rotation(__A ) __a = my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(__A ) return root class __A: def __init__( self ) -> None: '''simple docstring''' __a = None def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return get_height(self.root ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' print('''insert:''' + str(SCREAMING_SNAKE_CASE__ ) ) __a = insert_node(self.root , SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> None: '''simple docstring''' print('''delete:''' + str(SCREAMING_SNAKE_CASE__ ) ) if self.root is None: print('''Tree is empty!''' ) return __a = del_node(self.root , SCREAMING_SNAKE_CASE__ ) def __str__( self , ) -> str: # a level traversale, gives a more intuitive look on the tree '''simple docstring''' __a = '' __a = MyQueue() q.push(self.root ) __a = self.get_height() if layer == 0: return output __a = 0 while not q.is_empty(): __a = q.pop() __a = ' ' * int(math.pow(2 , layer - 1 ) ) output += space if node is None: output += "*" q.push(SCREAMING_SNAKE_CASE__ ) q.push(SCREAMING_SNAKE_CASE__ ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space __a = cnt + 1 for i in range(100 ): if cnt == math.pow(2 , SCREAMING_SNAKE_CASE__ ) - 1: __a = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def __lowerCAmelCase ( ) -> None: import doctest doctest.testmod() if __name__ == "__main__": _test() A : Dict = AVLtree() A : Union[str, Any] = list(range(1_0)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
6
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES UpperCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) UpperCAmelCase_ : str = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) UpperCAmelCase_ : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) UpperCAmelCase_ : int = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) UpperCAmelCase_ : Union[str, Any] = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) UpperCAmelCase_ : Union[str, Any] = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) UpperCAmelCase_ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) UpperCAmelCase_ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) UpperCAmelCase_ : Dict = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) UpperCAmelCase_ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) UpperCAmelCase_ : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) UpperCAmelCase_ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) UpperCAmelCase_ : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) UpperCAmelCase_ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) UpperCAmelCase_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) UpperCAmelCase_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) UpperCAmelCase_ : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[Any] = FLAX_MODEL_MAPPING UpperCAmelCase_ : Tuple = auto_class_update(FlaxAutoModel) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Any = FLAX_MODEL_FOR_PRETRAINING_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[Any] = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING UpperCAmelCase_ : Union[str, Any] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING UpperCAmelCase_ : Optional[int] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING UpperCAmelCase_ : str = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[str] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING UpperCAmelCase_ : Tuple = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Dict = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING UpperCAmelCase_ : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[int] = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING UpperCAmelCase_ : Dict = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING UpperCAmelCase_ : str = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[Any] = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING UpperCAmelCase_ : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[int] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING UpperCAmelCase_ : Union[str, Any] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
32
0
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated A__ : Tuple = collections.namedtuple('_Datasets', ['train', 'validation', 'test']) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ A__ : Any = 'https://storage.googleapis.com/cvdf-datasets/mnist/' def a ( lowerCamelCase_ ): '''simple docstring''' lowercase__ = numpy.dtype(numpy.uintaa ).newbyteorder('''>''' ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__A )[0] @deprecated(__A , '''Please use tf.data to implement this functionality.''' ) def a ( lowerCamelCase_ ): '''simple docstring''' print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=__A ) as bytestream: lowercase__ = _readaa(__A ) if magic != 2051: raise ValueError( '''Invalid magic number %d in MNIST image file: %s''' % (magic, f.name) ) lowercase__ = _readaa(__A ) lowercase__ = _readaa(__A ) lowercase__ = _readaa(__A ) lowercase__ = bytestream.read(rows * cols * num_images ) lowercase__ = numpy.frombuffer(__A , dtype=numpy.uinta ) lowercase__ = data.reshape(__A , __A , __A , 1 ) return data @deprecated(__A , '''Please use tf.one_hot on tensors.''' ) def a ( lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' lowercase__ = labels_dense.shape[0] lowercase__ = numpy.arange(__A ) * num_classes lowercase__ = numpy.zeros((num_labels, num_classes) ) lowercase__ = 1 return labels_one_hot @deprecated(__A , '''Please use tf.data to implement this functionality.''' ) def a ( lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=10 ): '''simple docstring''' print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=__A ) as bytestream: lowercase__ = _readaa(__A ) if magic != 2049: raise ValueError( '''Invalid magic number %d in MNIST label file: %s''' % (magic, f.name) ) lowercase__ = _readaa(__A ) lowercase__ = bytestream.read(__A ) lowercase__ = numpy.frombuffer(__A , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__A , __A ) return labels class _UpperCAmelCase : """simple docstring""" @deprecated( SCREAMING_SNAKE_CASE__, '''Please use alternatives such as official/mnist/_DataSet.py''' ''' from tensorflow/models.''', ) def __init__( self : List[str], lowerCamelCase : Optional[Any], lowerCamelCase : Dict, lowerCamelCase : List[Any]=False, lowerCamelCase : Tuple=False, lowerCamelCase : List[str]=dtypes.floataa, lowerCamelCase : Dict=True, lowerCamelCase : List[Any]=None, ): '''simple docstring''' lowercase__ = random_seed.get_seed(SCREAMING_SNAKE_CASE__ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase__ = dtypes.as_dtype(SCREAMING_SNAKE_CASE__ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError('''Invalid image dtype %r, expected uint8 or float32''' % dtype ) if fake_data: lowercase__ = 10_000 lowercase__ = one_hot else: assert ( images.shape[0] == labels.shape[0] ), F"""images.shape: {images.shape} labels.shape: {labels.shape}""" lowercase__ = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase__ = images.reshape( images.shape[0], images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase__ = images.astype(numpy.floataa ) lowercase__ = numpy.multiply(SCREAMING_SNAKE_CASE__, 1.0 / 255.0 ) lowercase__ = images lowercase__ = labels lowercase__ = 0 lowercase__ = 0 @property def lowercase__ ( self : Dict ): '''simple docstring''' return self._images @property def lowercase__ ( self : List[str] ): '''simple docstring''' return self._labels @property def lowercase__ ( self : Any ): '''simple docstring''' return self._num_examples @property def lowercase__ ( self : Union[str, Any] ): '''simple docstring''' return self._epochs_completed def lowercase__ ( self : Dict, lowerCamelCase : Union[str, Any], lowerCamelCase : Union[str, Any]=False, lowerCamelCase : Tuple=True ): '''simple docstring''' if fake_data: lowercase__ = [1] * 784 lowercase__ = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(SCREAMING_SNAKE_CASE__ )], [fake_label for _ in range(SCREAMING_SNAKE_CASE__ )], ) lowercase__ = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(SCREAMING_SNAKE_CASE__ ) lowercase__ = self.images[perma] lowercase__ = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase__ = self._num_examples - start lowercase__ = self._images[start : self._num_examples] lowercase__ = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase__ = numpy.arange(self._num_examples ) numpy.random.shuffle(SCREAMING_SNAKE_CASE__ ) lowercase__ = self.images[perm] lowercase__ = self.labels[perm] # Start next epoch lowercase__ = 0 lowercase__ = batch_size - rest_num_examples lowercase__ = self._index_in_epoch lowercase__ = self._images[start:end] lowercase__ = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part), axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part), axis=0 ), ) else: self._index_in_epoch += batch_size lowercase__ = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__A , '''Please write your own downloading logic.''' ) def a ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' if not gfile.Exists(__A ): gfile.MakeDirs(__A ) lowercase__ = os.path.join(__A , __A ) if not gfile.Exists(__A ): urllib.request.urlretrieve(__A , __A ) # noqa: S310 with gfile.GFile(__A ) as f: lowercase__ = f.size() print('''Successfully downloaded''' , __A , __A , '''bytes.''' ) return filepath @deprecated( __A , '''Please use alternatives such as:''' ''' tensorflow_datasets.load(\'mnist\')''' ) def a ( lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=False , lowerCamelCase_=dtypes.floataa , lowerCamelCase_=True , lowerCamelCase_=5000 , lowerCamelCase_=None , lowerCamelCase_=DEFAULT_SOURCE_URL , ): '''simple docstring''' if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__A , one_hot=__A , dtype=__A , seed=__A ) lowercase__ = fake() lowercase__ = fake() lowercase__ = fake() return _Datasets(train=__A , validation=__A , test=__A ) if not source_url: # empty string check lowercase__ = DEFAULT_SOURCE_URL lowercase__ = 'train-images-idx3-ubyte.gz' lowercase__ = 'train-labels-idx1-ubyte.gz' lowercase__ = 't10k-images-idx3-ubyte.gz' lowercase__ = 't10k-labels-idx1-ubyte.gz' lowercase__ = _maybe_download( __A , __A , source_url + train_images_file ) with gfile.Open(__A , '''rb''' ) as f: lowercase__ = _extract_images(__A ) lowercase__ = _maybe_download( __A , __A , source_url + train_labels_file ) with gfile.Open(__A , '''rb''' ) as f: lowercase__ = _extract_labels(__A , one_hot=__A ) lowercase__ = _maybe_download( __A , __A , source_url + test_images_file ) with gfile.Open(__A , '''rb''' ) as f: lowercase__ = _extract_images(__A ) lowercase__ = _maybe_download( __A , __A , source_url + test_labels_file ) with gfile.Open(__A , '''rb''' ) as f: lowercase__ = _extract_labels(__A , one_hot=__A ) if not 0 <= validation_size <= len(__A ): lowercase__ = ( 'Validation size should be between 0 and ' F"""{len(__A )}. Received: {validation_size}.""" ) raise ValueError(__A ) lowercase__ = train_images[:validation_size] lowercase__ = train_labels[:validation_size] lowercase__ = train_images[validation_size:] lowercase__ = train_labels[validation_size:] lowercase__ = {'dtype': dtype, 'reshape': reshape, 'seed': seed} lowercase__ = _DataSet(__A , __A , **__A ) lowercase__ = _DataSet(__A , __A , **__A ) lowercase__ = _DataSet(__A , __A , **__A ) return _Datasets(train=__A , validation=__A , test=__A )
207
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): snake_case__ : Any = GPTSanJapaneseTokenizer snake_case__ : Tuple = False snake_case__ : str = {'''do_clean_text''': False, '''add_prefix_space''': False} def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> str: super().setUp() # fmt: off a_ : Union[str, Any] = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on a_ : int = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 a_ : List[Any] = {'unk_token': '<unk>'} a_ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) a_ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(SCREAMING_SNAKE_CASE__ ) ) def SCREAMING_SNAKE_CASE ( self : List[str] , **SCREAMING_SNAKE_CASE__ : Tuple ) -> int: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> int: a_ : Optional[int] = 'こんにちは、世界。 \nこんばんは、㔺界。😀' a_ : List[str] = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : str , SCREAMING_SNAKE_CASE__ : int ) -> Dict: a_ , a_ : Union[str, Any] = self.get_input_output_texts(SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ ) a_ : Dict = tokenizer.decode(SCREAMING_SNAKE_CASE__ , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE__ ) return text, ids def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: pass # TODO add if relevant def SCREAMING_SNAKE_CASE ( self : Dict ) -> Any: pass # TODO add if relevant def SCREAMING_SNAKE_CASE ( self : Any ) -> Tuple: pass # TODO add if relevant def SCREAMING_SNAKE_CASE ( self : Dict ) -> List[str]: a_ : List[str] = self.get_tokenizer() # Testing tokenization a_ : List[Any] = 'こんにちは、世界。 こんばんは、㔺界。' a_ : Optional[int] = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] a_ : Dict = tokenizer.tokenize(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Testing conversion to ids without special tokens a_ : Tuple = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] a_ : List[Any] = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Testing conversion to ids with special tokens a_ : int = tokens + [tokenizer.unk_token] a_ : int = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 1_9] a_ : Tuple = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Dict: a_ : Union[str, Any] = self.get_tokenizer() # Testing tokenization a_ : Dict = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' a_ : List[Any] = 'こんにちは、、、、世界。こんばんは、、、、世界。' a_ : Any = tokenizer.encode(SCREAMING_SNAKE_CASE__ ) a_ : Optional[Any] = tokenizer.decode(SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : str ) -> Dict: a_ : Tuple = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization a_ : List[Any] = 'こんにちは、世界。' a_ : int = 'こんばんは、㔺界。😀' a_ : Dict = 'こんにちは、世界。こんばんは、世界。😀' a_ : Optional[int] = tokenizer.encode(prefix_text + input_text ) a_ : Any = tokenizer.encode('' , prefix_text=prefix_text + input_text ) a_ : Union[str, Any] = tokenizer.encode(SCREAMING_SNAKE_CASE__ , prefix_text=SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = tokenizer.decode(SCREAMING_SNAKE_CASE__ ) a_ : Tuple = tokenizer.decode(SCREAMING_SNAKE_CASE__ ) a_ : str = tokenizer.decode(SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Optional[int]: a_ : Tuple = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization a_ : str = 'こんにちは、世界。' a_ : List[str] = 'こんばんは、㔺界。😀' a_ : str = len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) - 2 a_ : Tuple = len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) - 2 a_ : Optional[Any] = [1] + [0] * (len_prefix + len_text + 1) a_ : Optional[Any] = [1] * (len_prefix + len_text + 1) + [0] a_ : Tuple = [1] + [1] * (len_prefix) + [0] * (len_text + 1) a_ : List[str] = tokenizer(prefix_text + input_text ).token_type_ids a_ : Union[str, Any] = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids a_ : Any = tokenizer(SCREAMING_SNAKE_CASE__ , prefix_text=SCREAMING_SNAKE_CASE__ ).token_type_ids self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: a_ : str = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) a_ : Optional[int] = tokenizer.encode('あンいワ' ) a_ : Dict = tokenizer.encode('' , prefix_text='あンいワ' ) a_ : Dict = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(SCREAMING_SNAKE_CASE__ ) , tokenizer.decode(SCREAMING_SNAKE_CASE__ ) ) self.assertEqual(tokenizer.decode(SCREAMING_SNAKE_CASE__ ) , tokenizer.decode(SCREAMING_SNAKE_CASE__ ) ) self.assertNotEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertNotEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[Any]: a_ : List[str] = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) a_ : Optional[Any] = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] a_ : List[str] = tokenizer(SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ ) a_ : Dict = tokenizer.batch_encode_plus(SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ ) # fmt: off a_ : List[Any] = [[3_5_9_9_3, 8_6_4_0, 2_5_9_4_8, 3_5_9_9_8, 3_0_6_4_7, 3_5_6_7_5, 3_5_9_9_9, 3_5_9_9_9], [3_5_9_9_3, 1_0_3_8_2, 9_8_6_8, 3_5_9_9_8, 3_0_6_4_6, 9_4_5_9, 3_0_6_4_6, 3_5_6_7_5]] a_ : Any = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] a_ : List[Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(x_token.token_type_ids , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(x_token.attention_mask , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(x_token_a.input_ids , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(x_token_a.token_type_ids , SCREAMING_SNAKE_CASE__ ) self.assertListEqual(x_token_a.attention_mask , SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: # tokenizer has no padding token pass
32
0
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCamelCase ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ : List[Any] =IFInpaintingSuperResolutionPipeline UpperCAmelCase_ : Dict =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} UpperCAmelCase_ : Dict =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) UpperCAmelCase_ : List[Any] =PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCAmelCase ( self ) -> Any: '''simple docstring''' return self._get_superresolution_dummy_components() def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase=0 ) -> Optional[Any]: '''simple docstring''' if str(SCREAMING_SNAKE_CASE__ ).startswith("mps" ): __snake_case : Any = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) else: __snake_case : Dict = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) __snake_case : int = floats_tensor((1, 3, 16, 16) , rng=random.Random(SCREAMING_SNAKE_CASE__ ) ).to(SCREAMING_SNAKE_CASE__ ) __snake_case : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(SCREAMING_SNAKE_CASE__ ) ).to(SCREAMING_SNAKE_CASE__ ) __snake_case : Optional[int] = floats_tensor((1, 3, 32, 32) , rng=random.Random(SCREAMING_SNAKE_CASE__ ) ).to(SCREAMING_SNAKE_CASE__ ) __snake_case : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' self._test_save_load_local() def UpperCAmelCase ( self ) -> Dict: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
326
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging UpperCAmelCase_ : Optional[int] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Union[str, Any] = ['''pixel_values'''] def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Optional[Dict[str, int]] = None , SCREAMING_SNAKE_CASE__ : PILImageResampling = PILImageResampling.BILINEAR , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Dict[str, int] = None , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Union[int, float] = 1 / 2_5_5 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None , **SCREAMING_SNAKE_CASE__ : Union[str, Any] , ) -> None: super().__init__(**SCREAMING_SNAKE_CASE__ ) a_ : str = size if size is not None else {'shortest_edge': 2_5_6} a_ : Any = get_size_dict(SCREAMING_SNAKE_CASE__ , default_to_square=SCREAMING_SNAKE_CASE__ ) a_ : Dict = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} a_ : Optional[int] = get_size_dict(SCREAMING_SNAKE_CASE__ ) a_ : List[str] = do_resize a_ : Dict = size a_ : Optional[Any] = resample a_ : Optional[int] = do_center_crop a_ : Dict = crop_size a_ : int = do_rescale a_ : int = rescale_factor a_ : Tuple = do_normalize a_ : int = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN a_ : Tuple = image_std if image_std is not None else IMAGENET_STANDARD_STD def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : Dict[str, int] , SCREAMING_SNAKE_CASE__ : PILImageResampling = PILImageResampling.BICUBIC , SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None , **SCREAMING_SNAKE_CASE__ : Union[str, Any] , ) -> np.ndarray: a_ : List[Any] = get_size_dict(SCREAMING_SNAKE_CASE__ , default_to_square=SCREAMING_SNAKE_CASE__ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) a_ : Tuple = get_resize_output_image_size(SCREAMING_SNAKE_CASE__ , size=size['shortest_edge'] , default_to_square=SCREAMING_SNAKE_CASE__ ) return resize(SCREAMING_SNAKE_CASE__ , size=SCREAMING_SNAKE_CASE__ , resample=SCREAMING_SNAKE_CASE__ , data_format=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : Dict[str, int] , SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None , **SCREAMING_SNAKE_CASE__ : Union[str, Any] , ) -> np.ndarray: a_ : str = get_size_dict(SCREAMING_SNAKE_CASE__ ) return center_crop(SCREAMING_SNAKE_CASE__ , size=(size['height'], size['width']) , data_format=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None , **SCREAMING_SNAKE_CASE__ : Optional[int] ) -> np.ndarray: return rescale(SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ , data_format=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : np.ndarray , SCREAMING_SNAKE_CASE__ : Union[float, List[float]] , SCREAMING_SNAKE_CASE__ : Union[float, List[float]] , SCREAMING_SNAKE_CASE__ : Optional[Union[str, ChannelDimension]] = None , **SCREAMING_SNAKE_CASE__ : List[str] , ) -> np.ndarray: return normalize(SCREAMING_SNAKE_CASE__ , mean=SCREAMING_SNAKE_CASE__ , std=SCREAMING_SNAKE_CASE__ , data_format=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : ImageInput , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Dict[str, int] = None , SCREAMING_SNAKE_CASE__ : PILImageResampling = None , SCREAMING_SNAKE_CASE__ : bool = None , SCREAMING_SNAKE_CASE__ : Dict[str, int] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[float] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = None , SCREAMING_SNAKE_CASE__ : Optional[Union[str, TensorType]] = None , SCREAMING_SNAKE_CASE__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE__ : Union[str, Any] , ) -> Union[str, Any]: a_ : List[str] = do_resize if do_resize is not None else self.do_resize a_ : Dict = size if size is not None else self.size a_ : Dict = get_size_dict(SCREAMING_SNAKE_CASE__ , default_to_square=SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = resample if resample is not None else self.resample a_ : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop a_ : int = crop_size if crop_size is not None else self.crop_size a_ : Optional[int] = get_size_dict(SCREAMING_SNAKE_CASE__ ) a_ : Dict = do_rescale if do_rescale is not None else self.do_rescale a_ : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor a_ : Any = do_normalize if do_normalize is not None else self.do_normalize a_ : str = image_mean if image_mean is not None else self.image_mean a_ : Dict = image_std if image_std is not None else self.image_std a_ : Optional[int] = make_list_of_images(SCREAMING_SNAKE_CASE__ ) if not valid_images(SCREAMING_SNAKE_CASE__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. a_ : Any = [to_numpy_array(SCREAMING_SNAKE_CASE__ ) for image in images] if do_resize: a_ : str = [self.resize(image=SCREAMING_SNAKE_CASE__ , size=SCREAMING_SNAKE_CASE__ , resample=SCREAMING_SNAKE_CASE__ ) for image in images] if do_center_crop: a_ : int = [self.center_crop(image=SCREAMING_SNAKE_CASE__ , size=SCREAMING_SNAKE_CASE__ ) for image in images] if do_rescale: a_ : Optional[Any] = [self.rescale(image=SCREAMING_SNAKE_CASE__ , scale=SCREAMING_SNAKE_CASE__ ) for image in images] if do_normalize: a_ : List[Any] = [self.normalize(image=SCREAMING_SNAKE_CASE__ , mean=SCREAMING_SNAKE_CASE__ , std=SCREAMING_SNAKE_CASE__ ) for image in images] a_ : Dict = [to_channel_dimension_format(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for image in images] a_ : Tuple = {'pixel_values': images} return BatchFeature(data=SCREAMING_SNAKE_CASE__ , tensor_type=SCREAMING_SNAKE_CASE__ )
32
0
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def __UpperCAmelCase ( __a : Optional[Any] ,__a : Union[str, Any] ,__a : int ,__a : Optional[int] ) -> Tuple: """simple docstring""" _a : Optional[Any] = s.rsplit(__A ,__A ) return new.join(__A ) def __UpperCAmelCase ( __a : List[Any] ) -> List[str]: """simple docstring""" return sum(param.float().sum() if '''encoder.embeddings''' not in key else 0 for key, param in state_dict.items() ) def __UpperCAmelCase ( __a : str ) -> int: """simple docstring""" _a : Tuple = {} _a : int = ['group_1', 'group_2', 'group_3', 'group_4'] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: _a : Optional[Any] = key.replace(F"""{group_key}.""" ,F"""{group_key}.group.""" ) if "res_path" in key: _a : Union[str, Any] = key.replace('''res_path.''' ,'''res_path.path.''' ) if key.endswith('''.w''' ): _a : Dict = rreplace(__A ,'''.w''' ,'''.weight''' ,1 ) if key.endswith('''.b''' ): _a : Optional[Any] = rreplace(__A ,'''.b''' ,'''.bias''' ,1 ) _a : Dict = value.float() return upgrade @torch.no_grad() def __UpperCAmelCase ( __a : List[str] ,__a : Optional[int] ,__a : Optional[int]=None ,__a : List[Any]=True ) -> List[Any]: """simple docstring""" from dall_e import Encoder _a : List[Any] = Encoder() if os.path.exists(__A ): _a : Tuple = torch.load(__A ) else: _a : List[Any] = torch.hub.load_state_dict_from_url(__A ) if isinstance(__A ,__A ): _a : Optional[int] = ckpt.state_dict() encoder.load_state_dict(__A ) if config_path is not None: _a : Optional[int] = FlavaImageCodebookConfig.from_pretrained(__A ) else: _a : Any = FlavaImageCodebookConfig() _a : int = FlavaImageCodebook(__A ).eval() _a : List[str] = encoder.state_dict() _a : Any = upgrade_state_dict(__A ) hf_model.load_state_dict(__A ) _a : int = hf_model.state_dict() _a : Optional[Any] = count_parameters(__A ) _a : Tuple = count_parameters(__A ) assert torch.allclose(__A ,__A ,atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__A ) else: return hf_state_dict if __name__ == "__main__": a__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to flava checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') a__ = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
235
def SCREAMING_SNAKE_CASE_ ( __A : list[int] , __A : str ) -> list[int]: """simple docstring""" a_ : Any = int(__A ) # Initialize Result a_ : Tuple = [] # Traverse through all denomination for denomination in reversed(__A ): # Find denominations while int(__A ) >= int(__A ): total_value -= int(__A ) answer.append(__A ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Union[str, Any] = '0' if ( input('Do you want to enter your denominations ? (yY/n): ').strip().lower() == "y" ): UpperCAmelCase_ : List[Any] = int(input('Enter the number of denominations you want to add: ').strip()) for i in range(0, n): denominations.append(int(input(F'Denomination {i}: ').strip())) UpperCAmelCase_ : str = input('Enter the change you want to make in Indian Currency: ').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase_ : List[Any] = [1, 2, 5, 10, 20, 50, 100, 500, 2000] UpperCAmelCase_ : str = input('Enter the change you want to make: ').strip() if int(value) == 0 or int(value) < 0: print('The total value cannot be zero or negative.') else: print(F'Following is minimal change for {value}: ') UpperCAmelCase_ : Optional[Any] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=' ')
32
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowerCAmelCase__ : Dict = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP', 'SpeechT5Config', 'SpeechT5HifiGanConfig', ], 'feature_extraction_speecht5': ['SpeechT5FeatureExtractor'], 'processing_speecht5': ['SpeechT5Processor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Tuple = ['SpeechT5Tokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[Any] = [ 'SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'SpeechT5ForSpeechToText', 'SpeechT5ForSpeechToSpeech', 'SpeechT5ForTextToSpeech', 'SpeechT5Model', 'SpeechT5PreTrainedModel', 'SpeechT5HifiGan', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
143
import flax.linen as nn import jax import jax.numpy as jnp class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : str ) -> int: a_ : Dict = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Optional[int]: a_ , a_ , a_ , a_ : Union[str, Any] = hidden_states.shape a_ : List[str] = jax.image.resize( SCREAMING_SNAKE_CASE__ , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) a_ : Any = self.conv(SCREAMING_SNAKE_CASE__ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: a_ : Optional[int] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Any , SCREAMING_SNAKE_CASE__ : int ) -> Optional[int]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a_ : str = self.conv(SCREAMING_SNAKE_CASE__ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : int = None snake_case__ : float = 0.0 snake_case__ : bool = None snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: a_ : List[str] = self.in_channels if self.out_channels is None else self.out_channels a_ : Optional[int] = nn.GroupNorm(num_groups=3_2 , epsilon=1E-5 ) a_ : Any = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a_ : Optional[int] = nn.Dense(SCREAMING_SNAKE_CASE__ , dtype=self.dtype ) a_ : Union[str, Any] = nn.GroupNorm(num_groups=3_2 , epsilon=1E-5 ) a_ : int = nn.Dropout(self.dropout_prob ) a_ : Optional[Any] = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a_ : List[str] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a_ : List[Any] = None if use_nin_shortcut: a_ : Union[str, Any] = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[Any]=True ) -> int: a_ : List[Any] = hidden_states a_ : Any = self.norma(SCREAMING_SNAKE_CASE__ ) a_ : Any = nn.swish(SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = self.conva(SCREAMING_SNAKE_CASE__ ) a_ : int = self.time_emb_proj(nn.swish(SCREAMING_SNAKE_CASE__ ) ) a_ : List[str] = jnp.expand_dims(jnp.expand_dims(SCREAMING_SNAKE_CASE__ , 1 ) , 1 ) a_ : Optional[int] = hidden_states + temb a_ : List[str] = self.norma(SCREAMING_SNAKE_CASE__ ) a_ : Tuple = nn.swish(SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = self.dropout(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = self.conva(SCREAMING_SNAKE_CASE__ ) if self.conv_shortcut is not None: a_ : List[str] = self.conv_shortcut(SCREAMING_SNAKE_CASE__ ) return hidden_states + residual
32
0
import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset __A ='bert-base-cased' __A ='google/pegasus-xsum' __A =[' Sam ate lunch today.', 'Sams lunch ingredients.'] __A =['A very interesting story about what I ate for lunch.', 'Avocado, celery, turkey, coffee'] __A ='patrickvonplaten/t5-tiny-random' __A ='sshleifer/bart-tiny-random' __A ='sshleifer/tiny-mbart' __A ='sshleifer/tiny-marian-en-de' def a ( _UpperCAmelCase : Path , _UpperCAmelCase : list ): '''simple docstring''' __UpperCAmelCase : List[str] = '\n'.join(__A ) Path(__A ).open('''w''' ).writelines(__A ) def a ( _UpperCAmelCase : Dict ): '''simple docstring''' for split in ["train", "val", "test"]: _dump_articles(os.path.join(__A , f'{split}.source' ) , __A ) _dump_articles(os.path.join(__A , f'{split}.target' ) , __A ) return tmp_dir class UpperCAmelCase__ ( lowercase__ ): '''simple docstring''' @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def snake_case__ ( self : int , a_ : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : Any = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Union[str, Any] = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __UpperCAmelCase : Any = max(len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) for a in ARTICLES ) __UpperCAmelCase : Tuple = max(len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) for a in SUMMARIES ) __UpperCAmelCase : Optional[Any] = 4 __UpperCAmelCase : str = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated __UpperCAmelCase : str = 'ro_RO', 'de_DE' # ignored for all but mbart, but never causes error. __UpperCAmelCase : str = SeqaSeqDataset( SCREAMING_SNAKE_CASE__ , data_dir=SCREAMING_SNAKE_CASE__ , type_path='''train''' , max_source_length=SCREAMING_SNAKE_CASE__ , max_target_length=SCREAMING_SNAKE_CASE__ , src_lang=SCREAMING_SNAKE_CASE__ , tgt_lang=SCREAMING_SNAKE_CASE__ , ) __UpperCAmelCase : Tuple = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place __UpperCAmelCase : Tuple = shift_tokens_right(batch['''labels'''] , tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def snake_case__ ( self : Optional[int] , a_ : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : str = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Any = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) __UpperCAmelCase : Any = max(len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) for a in ARTICLES ) __UpperCAmelCase : int = max(len(tokenizer.encode(SCREAMING_SNAKE_CASE__ ) ) for a in SUMMARIES ) __UpperCAmelCase : List[Any] = 4 __UpperCAmelCase : int = LegacySeqaSeqDataset( SCREAMING_SNAKE_CASE__ , data_dir=SCREAMING_SNAKE_CASE__ , type_path='''train''' , max_source_length=20 , max_target_length=SCREAMING_SNAKE_CASE__ , ) __UpperCAmelCase : Any = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 20 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def snake_case__ ( self : int ): '''simple docstring''' __UpperCAmelCase : int = AutoTokenizer.from_pretrained('''facebook/mbart-large-cc25''' ) __UpperCAmelCase : Tuple = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) __UpperCAmelCase : Union[str, Any] = tmp_dir.joinpath('''train.source''' ).open().readlines() __UpperCAmelCase : int = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1_28 , SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : List[str] = {x.name for x in tmp_dir.iterdir()} __UpperCAmelCase : Optional[Any] = {x.name for x in save_dir.iterdir()} __UpperCAmelCase : Optional[Any] = save_dir.joinpath('''train.source''' ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(SCREAMING_SNAKE_CASE__ ) < len(SCREAMING_SNAKE_CASE__ ) assert len(SCREAMING_SNAKE_CASE__ ) == 1 assert len(packed_examples[0] ) == sum(len(SCREAMING_SNAKE_CASE__ ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason='''This test requires fairseq''' ) def snake_case__ ( self : str ): '''simple docstring''' if not FAIRSEQ_AVAILABLE: return __UpperCAmelCase : Optional[int] = self._get_dataset(max_len=64 ) __UpperCAmelCase : Tuple = 64 __UpperCAmelCase : Any = ds.make_dynamic_sampler(SCREAMING_SNAKE_CASE__ , required_batch_size_multiple=SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : List[Any] = [len(SCREAMING_SNAKE_CASE__ ) for x in batch_sampler] assert len(set(SCREAMING_SNAKE_CASE__ ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ) # no dropped or added examples __UpperCAmelCase : Dict = DataLoader(SCREAMING_SNAKE_CASE__ , batch_sampler=SCREAMING_SNAKE_CASE__ , collate_fn=ds.collate_fn , num_workers=2 ) __UpperCAmelCase : Tuple = [] __UpperCAmelCase : Optional[int] = [] for batch in data_loader: __UpperCAmelCase : Any = batch['input_ids'].shape __UpperCAmelCase : Any = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple __UpperCAmelCase : List[Any] = np.product(batch['''input_ids'''].shape ) num_src_per_batch.append(SCREAMING_SNAKE_CASE__ ) if num_src_tokens > (max_tokens * 1.1): failures.append(SCREAMING_SNAKE_CASE__ ) assert num_src_per_batch[0] == max(SCREAMING_SNAKE_CASE__ ) if failures: raise AssertionError(F'too many tokens in {len(SCREAMING_SNAKE_CASE__ )} batches' ) def snake_case__ ( self : Dict ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = self._get_dataset(max_len=5_12 ) __UpperCAmelCase : Union[str, Any] = 2 __UpperCAmelCase : Tuple = ds.make_sortish_sampler(SCREAMING_SNAKE_CASE__ , shuffle=SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Tuple = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=SCREAMING_SNAKE_CASE__ , collate_fn=ds.collate_fn , num_workers=2 ) __UpperCAmelCase : Optional[Any] = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=SCREAMING_SNAKE_CASE__ , collate_fn=ds.collate_fn , num_workers=2 , sampler=SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Optional[Any] = tokenizer.pad_token_id def count_pad_tokens(a_ : Dict , a_ : Any="input_ids" ): return [batch[k].eq(SCREAMING_SNAKE_CASE__ ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(SCREAMING_SNAKE_CASE__ , k='''labels''' ) ) < sum(count_pad_tokens(SCREAMING_SNAKE_CASE__ , k='''labels''' ) ) assert sum(count_pad_tokens(SCREAMING_SNAKE_CASE__ ) ) < sum(count_pad_tokens(SCREAMING_SNAKE_CASE__ ) ) assert len(SCREAMING_SNAKE_CASE__ ) == len(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self : Optional[Any] , a_ : Any=10_00 , a_ : Any=1_28 ): '''simple docstring''' if os.getenv('''USE_REAL_DATA''' , SCREAMING_SNAKE_CASE__ ): __UpperCAmelCase : Optional[Any] = 'examples/seq2seq/wmt_en_ro' __UpperCAmelCase : Dict = max_len * 2 * 64 if not Path(SCREAMING_SNAKE_CASE__ ).joinpath('''train.len''' ).exists(): save_len_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: __UpperCAmelCase : Any = 'examples/seq2seq/test_data/wmt_en_ro' __UpperCAmelCase : Tuple = max_len * 4 save_len_file(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Any = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase : Optional[Any] = SeqaSeqDataset( SCREAMING_SNAKE_CASE__ , data_dir=SCREAMING_SNAKE_CASE__ , type_path='''train''' , max_source_length=SCREAMING_SNAKE_CASE__ , max_target_length=SCREAMING_SNAKE_CASE__ , n_obs=SCREAMING_SNAKE_CASE__ , ) return ds, max_tokens, tokenizer def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : Optional[int] = self._get_dataset() __UpperCAmelCase : Optional[Any] = set(DistributedSortishSampler(SCREAMING_SNAKE_CASE__ , 2_56 , num_replicas=2 , rank=0 , add_extra_examples=SCREAMING_SNAKE_CASE__ ) ) __UpperCAmelCase : List[Any] = set(DistributedSortishSampler(SCREAMING_SNAKE_CASE__ , 2_56 , num_replicas=2 , rank=1 , add_extra_examples=SCREAMING_SNAKE_CASE__ ) ) assert idsa.intersection(SCREAMING_SNAKE_CASE__ ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def snake_case__ ( self : Tuple , a_ : str ): '''simple docstring''' __UpperCAmelCase : str = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ , use_fast=SCREAMING_SNAKE_CASE__ ) if tok_name == MBART_TINY: __UpperCAmelCase : Dict = SeqaSeqDataset( SCREAMING_SNAKE_CASE__ , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , src_lang='''EN''' , tgt_lang='''FR''' , ) __UpperCAmelCase : Dict = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: __UpperCAmelCase : Any = SeqaSeqDataset( SCREAMING_SNAKE_CASE__ , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , ) __UpperCAmelCase : List[Any] = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(SCREAMING_SNAKE_CASE__ ) == 1 if tok_name == BART_TINY else len(SCREAMING_SNAKE_CASE__ ) == 0
226
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. UpperCAmelCase_ : Dict = {'LayoutLMv2Config', 'LayoutLMv3Config'} @is_pipeline_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): snake_case__ : List[str] = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING snake_case__ : Optional[Any] = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: snake_case__ : str = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: snake_case__ : List[Any] = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Tuple: a_ : List[Any] = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' ) a_ : int = text_classifier('This is great !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) a_ : Tuple = text_classifier('This is great !' , top_k=2 ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}] ) a_ : List[str] = text_classifier(['This is great !', 'This is bad'] , top_k=2 ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] , ) a_ : Tuple = text_classifier('This is great !' , top_k=1 ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) # Legacy behavior a_ : Union[str, Any] = text_classifier('This is great !' , return_all_scores=SCREAMING_SNAKE_CASE__ ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) a_ : List[str] = text_classifier('This is great !' , return_all_scores=SCREAMING_SNAKE_CASE__ ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [[{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}]] ) a_ : int = text_classifier(['This is great !', 'Something else'] , return_all_scores=SCREAMING_SNAKE_CASE__ ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] , ) a_ : str = text_classifier(['This is great !', 'Something else'] , return_all_scores=SCREAMING_SNAKE_CASE__ ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ {'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_0', 'score': 0.504}, ] , ) @require_torch def SCREAMING_SNAKE_CASE ( self : int ) -> Dict: import torch a_ : List[Any] = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' , device=torch.device('cpu' ) , ) a_ : Any = text_classifier('This is great !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) @require_tf def SCREAMING_SNAKE_CASE ( self : str ) -> Optional[Any]: a_ : List[str] = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='tf' ) a_ : Optional[int] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) @slow @require_torch def SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[int]: a_ : List[str] = pipeline('text-classification' ) a_ : Dict = text_classifier('This is great !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) a_ : Union[str, Any] = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) a_ : Tuple = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'POSITIVE', 'score': 0.988}] ) @slow @require_tf def SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[Any]: a_ : Dict = pipeline('text-classification' , framework='tf' ) a_ : Optional[Any] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) a_ : int = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) a_ : Optional[int] = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': 'POSITIVE', 'score': 0.988}] ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Any: a_ : Optional[Any] = TextClassificationPipeline(model=SCREAMING_SNAKE_CASE__ , tokenizer=SCREAMING_SNAKE_CASE__ ) return text_classifier, ["HuggingFace is in", "This is another test"] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Union[str, Any]: a_ : List[str] = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 a_ : Union[str, Any] = 'HuggingFace is in' a_ : int = text_classifier(SCREAMING_SNAKE_CASE__ ) self.assertEqual(nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}] ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) a_ : Union[str, Any] = ['HuggingFace is in ', 'Paris is in France'] a_ : int = text_classifier(SCREAMING_SNAKE_CASE__ ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}, {'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) self.assertTrue(outputs[1]['label'] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format a_ : List[Any] = text_classifier(SCREAMING_SNAKE_CASE__ , top_k=SCREAMING_SNAKE_CASE__ ) a_ : Dict = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [[{'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}] * N, [{'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}] * N] , ) a_ : int = {'text': 'HuggingFace is in ', 'text_pair': 'Paris is in France'} a_ : Optional[int] = text_classifier(SCREAMING_SNAKE_CASE__ ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , {'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )} , ) self.assertTrue(outputs['label'] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. a_ : Any = [['HuggingFace is in ', 'Paris is in France']] with self.assertRaises(SCREAMING_SNAKE_CASE__ ): text_classifier(SCREAMING_SNAKE_CASE__ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility a_ : Tuple = text_classifier([[['HuggingFace is in ', 'Paris is in France']]] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'label': ANY(SCREAMING_SNAKE_CASE__ ), 'score': ANY(SCREAMING_SNAKE_CASE__ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() )
32
0
def _lowerCAmelCase ( __lowerCAmelCase ) -> int: """simple docstring""" if not isinstance(__A , __A ): raise TypeError('''only integers accepted as input''' ) else: snake_case__ : Union[str, Any] = str(abs(__A ) ) snake_case__ : Optional[int] = [list(__A ) for char in range(len(__A ) )] for index in range(len(__A ) ): num_transpositions[index].pop(__A ) return max( int(''''''.join(list(__A ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('''doctest''').testmod()
230
import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig UpperCAmelCase_ : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase_ : str = 'T5Config' def SCREAMING_SNAKE_CASE_ ( __A : jnp.array , __A : int , __A : int ) -> jnp.ndarray: """simple docstring""" a_ : Dict = jnp.zeros_like(__A ) a_ : Dict = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) a_ : str = shifted_input_ids.at[:, 0].set(__A ) a_ : int = jnp.where(shifted_input_ids == -1_00 , __A , __A ) return shifted_input_ids class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : str = '''mt5''' snake_case__ : List[Any] = MTaConfig class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : str = '''mt5''' snake_case__ : List[str] = MTaConfig class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Any = '''mt5''' snake_case__ : Union[str, Any] = MTaConfig
32
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase : List[Any] = logging.get_logger(__name__) _lowercase : str = { 'facebook/deit-base-distilled-patch16-224': ( 'https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json' ), # See all DeiT models at https://huggingface.co/models?filter=deit } class _UpperCAmelCase ( lowercase__ ): a__ : Tuple = '''deit''' def __init__( self : Tuple , _lowercase : List[str]=7_68 , _lowercase : List[Any]=12 , _lowercase : List[str]=12 , _lowercase : Tuple=30_72 , _lowercase : Tuple="gelu" , _lowercase : Tuple=0.0 , _lowercase : Tuple=0.0 , _lowercase : Dict=0.02 , _lowercase : Tuple=1E-12 , _lowercase : Optional[int]=2_24 , _lowercase : Tuple=16 , _lowercase : Optional[Any]=3 , _lowercase : List[str]=True , _lowercase : Optional[Any]=16 , **_lowercase : List[Any] , ): super().__init__(**SCREAMING_SNAKE_CASE__ ) __UpperCAmelCase = hidden_size __UpperCAmelCase = num_hidden_layers __UpperCAmelCase = num_attention_heads __UpperCAmelCase = intermediate_size __UpperCAmelCase = hidden_act __UpperCAmelCase = hidden_dropout_prob __UpperCAmelCase = attention_probs_dropout_prob __UpperCAmelCase = initializer_range __UpperCAmelCase = layer_norm_eps __UpperCAmelCase = image_size __UpperCAmelCase = patch_size __UpperCAmelCase = num_channels __UpperCAmelCase = qkv_bias __UpperCAmelCase = encoder_stride class _UpperCAmelCase ( lowercase__ ): a__ : Any = version.parse("1.11" ) @property def a ( self : str ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def a ( self : int ): return 1E-4
332
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent UpperCAmelCase_ : Any = {'UserAgent': UserAgent().random} def SCREAMING_SNAKE_CASE_ ( __A : Optional[int] ) -> dict: """simple docstring""" a_ : Tuple = script.contents[0] a_ : int = json.loads(data[data.find('{"config"' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class SCREAMING_SNAKE_CASE__ : def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[Any]: a_ : Tuple = F"""https://www.instagram.com/{username}/""" a_ : Optional[Any] = self.get_json() def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> dict: a_ : Any = requests.get(self.url , headers=SCREAMING_SNAKE_CASE__ ).text a_ : Dict = BeautifulSoup(SCREAMING_SNAKE_CASE__ , 'html.parser' ).find_all('script' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : Union[str, Any] ) -> str: return F"""{self.__class__.__name__}('{self.username}')""" def __str__( self : Optional[int] ) -> str: return F"""{self.fullname} ({self.username}) is {self.biography}""" @property def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> str: return self.user_data["username"] @property def SCREAMING_SNAKE_CASE ( self : str ) -> str: return self.user_data["full_name"] @property def SCREAMING_SNAKE_CASE ( self : Any ) -> str: return self.user_data["biography"] @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> str: return self.user_data["business_email"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> str: return self.user_data["external_url"] @property def SCREAMING_SNAKE_CASE ( self : Dict ) -> int: return self.user_data["edge_followed_by"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Any ) -> int: return self.user_data["edge_follow"]["count"] @property def SCREAMING_SNAKE_CASE ( self : str ) -> int: return self.user_data["edge_owner_to_timeline_media"]["count"] @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> str: return self.user_data["profile_pic_url_hd"] @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> bool: return self.user_data["is_verified"] @property def SCREAMING_SNAKE_CASE ( self : Any ) -> bool: return self.user_data["is_private"] def SCREAMING_SNAKE_CASE_ ( __A : str = "github" ) -> None: """simple docstring""" import os if os.environ.get('CI' ): return # test failing on GitHub Actions a_ : int = InstagramUser(__A ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , __A ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 1_50 assert instagram_user.number_of_followers > 12_00_00 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('https://instagram.' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ : Union[str, Any] = InstagramUser('github') print(instagram_user) print(F'{instagram_user.number_of_posts = }') print(F'{instagram_user.number_of_followers = }') print(F'{instagram_user.number_of_followings = }') print(F'{instagram_user.email = }') print(F'{instagram_user.website = }') print(F'{instagram_user.profile_picture_url = }') print(F'{instagram_user.is_verified = }') print(F'{instagram_user.is_private = }')
32
0
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __A : int = logging.get_logger(__name__) @add_end_docstrings(lowercase__) class _SCREAMING_SNAKE_CASE ( lowercase__): def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )-> List[Any]: super().__init__(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _snake_case ( self , _SCREAMING_SNAKE_CASE=None )-> List[Any]: lowerCamelCase_ ={} if top_k is not None: lowerCamelCase_ =top_k return {}, {}, postprocess_params def __call__( self , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )-> Tuple: return super().__call__(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Union[str, Any]: lowerCamelCase_ =load_image(SCREAMING_SNAKE_CASE__ ) lowerCamelCase_ =self.image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors=self.framework ) return model_inputs def _snake_case ( self , _SCREAMING_SNAKE_CASE )-> Dict: lowerCamelCase_ =self.model(**SCREAMING_SNAKE_CASE__ ) return model_outputs def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=5 )-> Any: if top_k > self.model.config.num_labels: lowerCamelCase_ =self.model.config.num_labels if self.framework == "pt": lowerCamelCase_ =model_outputs.logits.softmax(-1 )[0] lowerCamelCase_ =probs.topk(SCREAMING_SNAKE_CASE__ ) elif self.framework == "tf": lowerCamelCase_ =stable_softmax(model_outputs.logits , axis=-1 )[0] lowerCamelCase_ =tf.math.top_k(SCREAMING_SNAKE_CASE__ , k=SCREAMING_SNAKE_CASE__ ) lowerCamelCase_ =topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'Unsupported framework: {self.framework}' ) lowerCamelCase_ =scores.tolist() lowerCamelCase_ =ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ )]
154
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Tuple = ['''image_processor''', '''tokenizer'''] snake_case__ : Union[str, Any] = '''CLIPImageProcessor''' snake_case__ : Dict = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , SCREAMING_SNAKE_CASE__ : List[str]=None , **SCREAMING_SNAKE_CASE__ : int ) -> Any: a_ : List[Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , SCREAMING_SNAKE_CASE__ , ) a_ : Tuple = kwargs.pop('feature_extractor' ) a_ : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __call__( self : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , **SCREAMING_SNAKE_CASE__ : str ) -> Optional[Any]: if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: a_ : List[str] = self.tokenizer(SCREAMING_SNAKE_CASE__ , return_tensors=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if images is not None: a_ : Dict = self.image_processor(SCREAMING_SNAKE_CASE__ , return_tensors=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) if text is not None and images is not None: a_ : Dict = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**SCREAMING_SNAKE_CASE__ ) , tensor_type=SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Any , *SCREAMING_SNAKE_CASE__ : List[Any] , **SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[Any]: return self.tokenizer.batch_decode(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , *SCREAMING_SNAKE_CASE__ : List[Any] , **SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[Any]: return self.tokenizer.decode(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @property def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Any: a_ : str = self.tokenizer.model_input_names a_ : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def SCREAMING_SNAKE_CASE ( self : str ) -> str: warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , SCREAMING_SNAKE_CASE__ , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE ( self : str ) -> Tuple: warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , SCREAMING_SNAKE_CASE__ , ) return self.image_processor
32
0
'''simple docstring''' import argparse import struct import unittest class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase ) -> None: lowercase__ : Tuple = data # Initialize hash values lowercase__ : Tuple = [ 0X6A09E667, 0XBB67AE85, 0X3C6EF372, 0XA54FF53A, 0X510E527F, 0X9B05688C, 0X1F83D9AB, 0X5BE0CD19, ] # Initialize round constants lowercase__ : Tuple = [ 0X428A2F98, 0X71374491, 0XB5C0FBCF, 0XE9B5DBA5, 0X3956C25B, 0X59F111F1, 0X923F82A4, 0XAB1C5ED5, 0XD807AA98, 0X12835B01, 0X243185BE, 0X550C7DC3, 0X72BE5D74, 0X80DEB1FE, 0X9BDC06A7, 0XC19BF174, 0XE49B69C1, 0XEFBE4786, 0X0FC19DC6, 0X240CA1CC, 0X2DE92C6F, 0X4A7484AA, 0X5CB0A9DC, 0X76F988DA, 0X983E5152, 0XA831C66D, 0XB00327C8, 0XBF597FC7, 0XC6E00BF3, 0XD5A79147, 0X06CA6351, 0X14292967, 0X27B70A85, 0X2E1B2138, 0X4D2C6DFC, 0X53380D13, 0X650A7354, 0X766A0ABB, 0X81C2C92E, 0X92722C85, 0XA2BFE8A1, 0XA81A664B, 0XC24B8B70, 0XC76C51A3, 0XD192E819, 0XD6990624, 0XF40E3585, 0X106AA070, 0X19A4C116, 0X1E376C08, 0X2748774C, 0X34B0BCB5, 0X391C0CB3, 0X4ED8AA4A, 0X5B9CCA4F, 0X682E6FF3, 0X748F82EE, 0X78A5636F, 0X84C87814, 0X8CC70208, 0X90BEFFFA, 0XA4506CEB, 0XBEF9A3F7, 0XC67178F2, ] lowercase__ : int = self.preprocessing(self.data ) self.final_hash() @staticmethod def _lowerCAmelCase( __lowerCAmelCase ) -> bytes: lowercase__ : Any = b'\x80' + (b'\x00' * (63 - (len(SCREAMING_SNAKE_CASE__ ) + 8) % 64)) lowercase__ : List[str] = struct.pack('''>Q''' , (len(SCREAMING_SNAKE_CASE__ ) * 8) ) return data + padding + big_endian_integer def _lowerCAmelCase( self ) -> None: # Convert into blocks of 64 bytes lowercase__ : int = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers lowercase__ : List[str] = list(struct.unpack('''>16L''' , SCREAMING_SNAKE_CASE__ ) ) # add 48 0-ed integers words += [0] * 48 lowercase__ : Tuple = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array lowercase__ : Optional[int] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) lowercase__ : List[str] = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) lowercase__ : Union[str, Any] = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X100000000 # Compression lowercase__ : Any = self.ror(SCREAMING_SNAKE_CASE__ , 6 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 11 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 25 ) lowercase__ : List[str] = (e & f) ^ ((~e & 0XFFFFFFFF) & g) lowercase__ : Tuple = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X100000000 lowercase__ : int = self.ror(SCREAMING_SNAKE_CASE__ , 2 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 13 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 22 ) lowercase__ : Optional[Any] = (a & b) ^ (a & c) ^ (b & c) lowercase__ : Tuple = (sa + maj) % 0X100000000 lowercase__ : List[str] = ( g, f, e, ((d + tempa) % 0X100000000), c, b, a, ((tempa + tempa) % 0X100000000), ) lowercase__ : Tuple = [a, b, c, d, e, f, g, h] # Modify final values lowercase__ : Dict = [ ((element + mutated_hash_values[index]) % 0X100000000) for index, element in enumerate(self.hashes ) ] lowercase__ : Any = ''.join([hex(SCREAMING_SNAKE_CASE__ )[2:].zfill(8 ) for value in self.hashes] ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> int: return 0XFFFFFFFF & (value << (32 - rotations)) | (value >> rotations) class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase( self ) -> None: import hashlib lowercase__ : Union[str, Any] = bytes('''Test String''' , '''utf-8''' ) self.assertEqual(SHAaaa(SCREAMING_SNAKE_CASE__ ).hash , hashlib.shaaaa(SCREAMING_SNAKE_CASE__ ).hexdigest() ) def __UpperCamelCase ( ): import doctest doctest.testmod() lowercase__ : Dict = argparse.ArgumentParser() parser.add_argument( '''-s''' , '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument( '''-f''' , '''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) lowercase__ : Optional[int] = parser.parse_args() lowercase__ : List[Any] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: lowercase__ : Optional[Any] = f.read() else: lowercase__ : Optional[Any] = bytes(__A , '''utf-8''' ) print(SHAaaa(__A ).hash ) if __name__ == "__main__": main()
198
from __future__ import annotations UpperCAmelCase_ : Tuple = [] def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int , __A : int ) -> bool: """simple docstring""" for i in range(len(__A ) ): if board[row][i] == 1: return False for i in range(len(__A ) ): if board[i][column] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , len(__A ) ) ): if board[i][j] == 1: return False return True def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int ) -> bool: """simple docstring""" if row >= len(__A ): solution.append(__A ) printboard(__A ) print() return True for i in range(len(__A ) ): if is_safe(__A , __A , __A ): a_ : Any = 1 solve(__A , row + 1 ) a_ : Tuple = 0 return False def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] ) -> None: """simple docstring""" for i in range(len(__A ) ): for j in range(len(__A ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) UpperCAmelCase_ : List[str] = 8 UpperCAmelCase_ : str = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
32
0
"""simple docstring""" def __A ( a_ :int = 10_00) -> int: __a : Union[str, Any] = 2**power __a : Tuple = str(__A) __a : int = list(__A) __a : Optional[Any] = 0 for i in list_num: sum_of_num += int(__A) return sum_of_num if __name__ == "__main__": A = int(input('''Enter the power of 2: ''').strip()) print('''2 ^ ''', power, ''' = ''', 2**power) A = solution(power) print('''Sum of the digits is: ''', result)
160
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def SCREAMING_SNAKE_CASE_ ( ) -> Any: """simple docstring""" a_ : Optional[Any] = HfArgumentParser(__A ) a_ : Optional[int] = parser.parse_args_into_dataclasses()[0] a_ : List[Any] = TensorFlowBenchmark(args=__A ) try: a_ : List[str] = parser.parse_args_into_dataclasses()[0] except ValueError as e: a_ : Dict = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' a_ : Dict = ' '.join(str(__A ).split(' ' )[:-1] ) a_ : int = '' a_ : int = eval(str(__A ).split(' ' )[-1] ) a_ : Any = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(__A ) if len(__A ) > 0: a_ : str = full_error_msg + begin_error_msg + str(__A ) raise ValueError(__A ) benchmark.run() if __name__ == "__main__": main()
32
0
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class __A: def __init__( self , _snake_case , ) -> List[str]: '''simple docstring''' __a = parent __a = 13 __a = 7 __a = 30 __a = self.seq_length + self.mem_len __a = 15 __a = True __a = True __a = 99 __a = [10, 50, 80] __a = 32 __a = 32 __a = 4 __a = 8 __a = 128 __a = 2 __a = 2 __a = None __a = 1 __a = 0 __a = 3 __a = self.vocab_size - 1 __a = 0.01 def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' random.seed(self.seed ) tf.random.set_seed(self.seed ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> List[Any]: '''simple docstring''' __a = TFTransfoXLModel(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = {'input_ids': input_ids_a, 'mems': mems_a} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> str: '''simple docstring''' __a = TFTransfoXLLMHeadModel(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = {'input_ids': input_ids_a, 'labels': lm_labels} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = model([input_ids_a, mems_a] ).to_tuple() __a = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> Dict: '''simple docstring''' __a = TFTransfoXLForSequenceClassification(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self ) -> Union[str, Any]: '''simple docstring''' __a = self.prepare_config_and_inputs() (__a) = config_and_inputs __a = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class __A( lowercase__ , lowercase__ , unittest.TestCase ): snake_case_ = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) snake_case_ = () if is_tf_available() else () snake_case_ = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) -> str: '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' __a = TFTransfoXLModelTester(self ) __a = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , d_embed=37 ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' self.model_tester.set_seed() __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' self.model_tester.set_seed() __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs_for_common() __a = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: __a = model_class(SCREAMING_SNAKE_CASE__ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: __a = model.get_output_embeddings() assert isinstance(SCREAMING_SNAKE_CASE__ , tf.keras.layers.Layer ) __a = model.get_bias() assert name is None else: __a = model.get_output_embeddings() assert x is None __a = model.get_bias() assert name is None def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' pass @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = TFTransfoXLModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @unittest.skip(reason='''This model doesn\'t play well with fit() due to not returning a single loss.''' ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' pass @require_tf class __A( unittest.TestCase ): @unittest.skip('''Skip test until #12651 is resolved.''' ) @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' __a = TFTransfoXLLMHeadModel.from_pretrained('''transfo-xl-wt103''' ) # fmt: off __a = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off __a = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> __a = model.generate(SCREAMING_SNAKE_CASE__ , max_length=200 , do_sample=SCREAMING_SNAKE_CASE__ ) self.assertListEqual(output_ids[0].numpy().tolist() , SCREAMING_SNAKE_CASE__ )
6
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): snake_case__ : Optional[Any] = TextToVideoSDPipeline snake_case__ : Optional[int] = TEXT_TO_IMAGE_PARAMS snake_case__ : str = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. snake_case__ : Optional[Any] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: torch.manual_seed(0 ) a_ : Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4, 6_4, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=3_2 , attention_head_dim=4 , ) a_ : int = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=SCREAMING_SNAKE_CASE__ , set_alpha_to_one=SCREAMING_SNAKE_CASE__ , ) torch.manual_seed(0 ) a_ : int = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) a_ : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=5_1_2 , ) a_ : Dict = CLIPTextModel(SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) a_ : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : List[Any]=0 ) -> List[str]: if str(SCREAMING_SNAKE_CASE__ ).startswith('mps' ): a_ : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) else: a_ : Optional[Any] = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) a_ : int = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: a_ : List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator a_ : Dict = self.get_dummy_components() a_ : str = TextToVideoSDPipeline(**SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = sd_pipe.to(SCREAMING_SNAKE_CASE__ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Tuple = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) a_ : Dict = 'np' a_ : Dict = sd_pipe(**SCREAMING_SNAKE_CASE__ ).frames a_ : int = frames[0][-3:, -3:, -1] assert frames[0].shape == (6_4, 6_4, 3) a_ : Union[str, Any] = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE ( self : Any ) -> List[Any]: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=SCREAMING_SNAKE_CASE__ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def SCREAMING_SNAKE_CASE ( self : Any ) -> str: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=SCREAMING_SNAKE_CASE__ , expected_max_diff=1E-2 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: pass def SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: return super().test_progress_bar() @slow @skip_mps class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: a_ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy' ) a_ : Any = TextToVideoSDPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b' ) a_ : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) a_ : Optional[Any] = pipe.to('cuda' ) a_ : Any = 'Spiderman is surfing' a_ : List[Any] = torch.Generator(device='cpu' ).manual_seed(0 ) a_ : Optional[Any] = pipe(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2_5 , output_type='pt' ).frames a_ : str = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def SCREAMING_SNAKE_CASE ( self : Any ) -> Any: a_ : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy' ) a_ : Tuple = TextToVideoSDPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b' ) a_ : Tuple = pipe.to('cuda' ) a_ : Any = 'Spiderman is surfing' a_ : List[str] = torch.Generator(device='cpu' ).manual_seed(0 ) a_ : List[Any] = pipe(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type='pt' ).frames a_ : List[str] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
32
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A__ : Optional[Any] = logging.get_logger(__name__) A__ : List[Any] = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class _UpperCAmelCase ( lowercase__ ,lowercase__ ): """simple docstring""" lowercase__ = '''bit''' lowercase__ = ['''preactivation''', '''bottleneck'''] lowercase__ = ['''SAME''', '''VALID'''] def __init__( self : Dict, lowerCamelCase : List[Any]=3, lowerCamelCase : Optional[int]=64, lowerCamelCase : Optional[int]=[256, 512, 1_024, 2_048], lowerCamelCase : Optional[Any]=[3, 4, 6, 3], lowerCamelCase : Optional[Any]="preactivation", lowerCamelCase : Tuple="relu", lowerCamelCase : List[str]=None, lowerCamelCase : Dict=32, lowerCamelCase : Tuple=0.0, lowerCamelCase : Any=False, lowerCamelCase : Tuple=32, lowerCamelCase : str=1, lowerCamelCase : Union[str, Any]=None, lowerCamelCase : Union[str, Any]=None, **lowerCamelCase : str, ): '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: lowercase__ = global_padding.upper() else: raise ValueError(F"""Padding strategy {global_padding} not supported""" ) lowercase__ = num_channels lowercase__ = embedding_size lowercase__ = hidden_sizes lowercase__ = depths lowercase__ = layer_type lowercase__ = hidden_act lowercase__ = global_padding lowercase__ = num_groups lowercase__ = drop_path_rate lowercase__ = embedding_dynamic_padding lowercase__ = output_stride lowercase__ = width_factor lowercase__ = ['stem'] + [F"""stage{idx}""" for idx in range(1, len(SCREAMING_SNAKE_CASE__ ) + 1 )] lowercase__ = get_aligned_output_features_output_indices( out_features=SCREAMING_SNAKE_CASE__, out_indices=SCREAMING_SNAKE_CASE__, stage_names=self.stage_names )
207
import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): # TODO: is there an appropriate internal test set? snake_case__ : Any = '''ssube/stable-diffusion-x4-upscaler-onnx''' def SCREAMING_SNAKE_CASE ( self : List[str] , SCREAMING_SNAKE_CASE__ : int=0 ) -> Tuple: a_ : Union[str, Any] = floats_tensor((1, 3, 1_2_8, 1_2_8) , rng=random.Random(SCREAMING_SNAKE_CASE__ ) ) a_ : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Dict: a_ : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = self.get_dummy_inputs() a_ : int = pipe(**SCREAMING_SNAKE_CASE__ ).images a_ : Tuple = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_1_2, 5_1_2, 3) a_ : List[Any] = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: a_ : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) a_ : int = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : List[str] = self.get_dummy_inputs() a_ : List[str] = pipe(**SCREAMING_SNAKE_CASE__ ).images a_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a_ : str = np.array( [0.6898892, 0.59240556, 0.52499527, 0.58866215, 0.52258235, 0.52572715, 0.62414473, 0.6174387, 0.6214964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def SCREAMING_SNAKE_CASE ( self : Any ) -> Dict: a_ : Optional[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) a_ : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = self.get_dummy_inputs() a_ : Dict = pipe(**SCREAMING_SNAKE_CASE__ ).images a_ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a_ : Optional[Any] = np.array( [0.7659278, 0.76437664, 0.75579107, 0.7691116, 0.77666986, 0.7727672, 0.7758664, 0.7812226, 0.76942515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: a_ : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) a_ : int = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Optional[Any] = self.get_dummy_inputs() a_ : Dict = pipe(**SCREAMING_SNAKE_CASE__ ).images a_ : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a_ : int = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def SCREAMING_SNAKE_CASE ( self : str ) -> Optional[Any]: a_ : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) a_ : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = self.get_dummy_inputs() a_ : List[str] = pipe(**SCREAMING_SNAKE_CASE__ ).images a_ : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a_ : Union[str, Any] = np.array( [0.77424496, 0.773601, 0.7645288, 0.7769598, 0.7772739, 0.7738688, 0.78187233, 0.77879584, 0.767043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @property def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Union[str, Any]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[str]: a_ : List[str] = ort.SessionOptions() a_ : int = False return options def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Tuple: a_ : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) a_ : int = init_image.resize((1_2_8, 1_2_8) ) # using the PNDM scheduler by default a_ : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Tuple = 'A fantasy landscape, trending on artstation' a_ : str = torch.manual_seed(0 ) a_ : List[str] = pipe( prompt=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , guidance_scale=7.5 , num_inference_steps=1_0 , generator=SCREAMING_SNAKE_CASE__ , output_type='np' , ) a_ : Dict = output.images a_ : Any = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) a_ : str = np.array([0.4883, 0.4947, 0.4980, 0.4975, 0.4982, 0.4980, 0.5000, 0.5006, 0.4972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[str]: a_ : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) a_ : List[str] = init_image.resize((1_2_8, 1_2_8) ) a_ : Dict = LMSDiscreteScheduler.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , subfolder='scheduler' ) a_ : Union[str, Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , scheduler=SCREAMING_SNAKE_CASE__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Any = 'A fantasy landscape, trending on artstation' a_ : Tuple = torch.manual_seed(0 ) a_ : Optional[Any] = pipe( prompt=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , guidance_scale=7.5 , num_inference_steps=2_0 , generator=SCREAMING_SNAKE_CASE__ , output_type='np' , ) a_ : str = output.images a_ : List[Any] = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) a_ : Tuple = np.array( [0.50173753, 0.50223356, 0.502039, 0.50233036, 0.5023725, 0.5022601, 0.5018758, 0.50234085, 0.50241566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
32
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule __A : Any = {'''tokenization_byt5''': ['''ByT5Tokenizer''']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin __A : Dict = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class _UpperCAmelCase ( unittest.TestCase , _A ): def A ( self : List[Any] ) -> Dict: lowercase_ : Optional[int] = load_tool('''text-question-answering''' ) self.tool.setup() lowercase_ : Union[str, Any] = load_tool('''text-question-answering''' , remote=A ) def A ( self : Any ) -> List[str]: lowercase_ : Union[str, Any] = self.tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : str ) -> List[str]: lowercase_ : int = self.remote_tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[Any] ) -> int: lowercase_ : Optional[Any] = self.tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[str] ) -> Optional[int]: lowercase_ : int = self.remote_tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' )
33
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __A : List[Any] = logging.get_logger(__name__) __A : List[Any] = { '''microsoft/resnet-50''': '''https://huggingface.co/microsoft/resnet-50/blob/main/config.json''', } class _UpperCAmelCase ( _A , _A ): SCREAMING_SNAKE_CASE_ : List[Any] = "resnet" SCREAMING_SNAKE_CASE_ : Tuple = ["basic", "bottleneck"] def __init__( self : Any , A : Tuple=3 , A : str=64 , A : Tuple=[2_56, 5_12, 10_24, 20_48] , A : List[Any]=[3, 4, 6, 3] , A : Union[str, Any]="bottleneck" , A : int="relu" , A : List[Any]=False , A : Tuple=None , A : int=None , **A : List[str] , ) -> List[Any]: super().__init__(**A ) if layer_type not in self.layer_types: raise ValueError(F'''layer_type={layer_type} is not one of {','.join(self.layer_types )}''' ) lowercase_ : List[Any] = num_channels lowercase_ : Tuple = embedding_size lowercase_ : Dict = hidden_sizes lowercase_ : Tuple = depths lowercase_ : Optional[int] = layer_type lowercase_ : str = hidden_act lowercase_ : Dict = downsample_in_first_stage lowercase_ : str = ['''stem'''] + [F'''stage{idx}''' for idx in range(1 , len(A ) + 1 )] lowercase_ , lowercase_ : List[str] = get_aligned_output_features_output_indices( out_features=A , out_indices=A , stage_names=self.stage_names ) class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : str = version.parse("1.11" ) @property def A ( self : Any ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def A ( self : Union[str, Any] ) -> float: return 1e-3
33
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _UpperCAmelCase ( TensorFormatter[Mapping, "torch.Tensor", Mapping] ): def __init__( self : Any , A : int=None , **A : str ) -> Union[str, Any]: super().__init__(features=A ) lowercase_ : Union[str, Any] = torch_tensor_kwargs import torch # noqa import torch at initialization def A ( self : Dict , A : int ) -> List[Any]: import torch if isinstance(A , A ) and column: if all( isinstance(A , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(A ) return column def A ( self : int , A : Any ) -> Optional[Any]: import torch if isinstance(A , (str, bytes, type(A )) ): return value elif isinstance(A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() lowercase_ : Any = {} if isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): lowercase_ : Any = {'''dtype''': torch.intaa} elif isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): lowercase_ : Dict = {'''dtype''': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A , PIL.Image.Image ): lowercase_ : Dict = np.asarray(A ) return torch.tensor(A , **{**default_dtype, **self.torch_tensor_kwargs} ) def A ( self : Union[str, Any] , A : Optional[int] ) -> str: import torch # support for torch, tf, jax etc. if hasattr(A , '''__array__''' ) and not isinstance(A , torch.Tensor ): lowercase_ : Optional[int] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) elif isinstance(A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) return self._tensorize(A ) def A ( self : Dict , A : dict ) -> Tuple: return map_nested(self._recursive_tensorize , A , map_list=A ) def A ( self : str , A : pa.Table ) -> Mapping: lowercase_ : Optional[Any] = self.numpy_arrow_extractor().extract_row(A ) lowercase_ : str = self.python_features_decoder.decode_row(A ) return self.recursive_tensorize(A ) def A ( self : List[Any] , A : pa.Table ) -> "torch.Tensor": lowercase_ : List[str] = self.numpy_arrow_extractor().extract_column(A ) lowercase_ : str = self.python_features_decoder.decode_column(A , pa_table.column_names[0] ) lowercase_ : Optional[int] = self.recursive_tensorize(A ) lowercase_ : Any = self._consolidate(A ) return column def A ( self : List[str] , A : pa.Table ) -> Mapping: lowercase_ : Optional[int] = self.numpy_arrow_extractor().extract_batch(A ) lowercase_ : int = self.python_features_decoder.decode_batch(A ) lowercase_ : Dict = self.recursive_tensorize(A ) for column_name in batch: lowercase_ : Optional[Any] = self._consolidate(batch[column_name] ) return batch
33
1
"""simple docstring""" def lowercase ( __snake_case : int , __snake_case : int , __snake_case : int ): lowercase_ : Union[str, Any] = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def lowercase ( ): print(sum_of_series(1 , 1 , 1_0 ) ) if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
33
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A : str = logging.get_logger(__name__) __A : List[str] = { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json''', # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : List[str] = "convbert" def __init__( self : int , A : Optional[int]=3_05_22 , A : Dict=7_68 , A : Dict=12 , A : Optional[int]=12 , A : Any=30_72 , A : Union[str, Any]="gelu" , A : Any=0.1 , A : Dict=0.1 , A : int=5_12 , A : int=2 , A : int=0.02 , A : Tuple=1e-12 , A : int=1 , A : int=0 , A : str=2 , A : Optional[int]=7_68 , A : List[Any]=2 , A : Optional[int]=9 , A : Optional[int]=1 , A : str=None , **A : List[Any] , ) -> Optional[Any]: super().__init__( pad_token_id=A , bos_token_id=A , eos_token_id=A , **A , ) lowercase_ : List[Any] = vocab_size lowercase_ : str = hidden_size lowercase_ : List[Any] = num_hidden_layers lowercase_ : Any = num_attention_heads lowercase_ : Dict = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : str = hidden_dropout_prob lowercase_ : Dict = attention_probs_dropout_prob lowercase_ : List[Any] = max_position_embeddings lowercase_ : Any = type_vocab_size lowercase_ : Optional[int] = initializer_range lowercase_ : Optional[Any] = layer_norm_eps lowercase_ : Dict = embedding_size lowercase_ : Any = head_ratio lowercase_ : Union[str, Any] = conv_kernel_size lowercase_ : int = num_groups lowercase_ : Dict = classifier_dropout class _UpperCAmelCase ( _A ): @property def A ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowercase_ : Any = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase_ : Tuple = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
33
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __A : Union[str, Any] = logging.get_logger(__name__) # General docstring __A : Tuple = '''MobileNetV1Config''' # Base docstring __A : Union[str, Any] = '''google/mobilenet_v1_1.0_224''' __A : Union[str, Any] = [1, 1_024, 7, 7] # Image classification docstring __A : Optional[Any] = '''google/mobilenet_v1_1.0_224''' __A : List[Any] = '''tabby, tabby cat''' __A : Union[str, Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=None ): lowercase_ : str = {} if isinstance(__snake_case , __snake_case ): lowercase_ : Union[str, Any] = model.mobilenet_va else: lowercase_ : Optional[Any] = model lowercase_ : Union[str, Any] = '''MobilenetV1/Conv2d_0/''' lowercase_ : Union[str, Any] = backbone.conv_stem.convolution.weight lowercase_ : Optional[Any] = backbone.conv_stem.normalization.bias lowercase_ : Union[str, Any] = backbone.conv_stem.normalization.weight lowercase_ : Any = backbone.conv_stem.normalization.running_mean lowercase_ : int = backbone.conv_stem.normalization.running_var for i in range(1_3 ): lowercase_ : Optional[int] = i + 1 lowercase_ : Union[str, Any] = i * 2 lowercase_ : Optional[Any] = backbone.layer[pt_index] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' lowercase_ : str = pointer.convolution.weight lowercase_ : int = pointer.normalization.bias lowercase_ : Any = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Union[str, Any] = pointer.normalization.running_var lowercase_ : Any = backbone.layer[pt_index + 1] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' lowercase_ : int = pointer.convolution.weight lowercase_ : str = pointer.normalization.bias lowercase_ : Tuple = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Any = pointer.normalization.running_var if isinstance(__snake_case , __snake_case ): lowercase_ : Optional[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase_ : Any = model.classifier.weight lowercase_ : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase_ : Tuple = tf.train.list_variables(__snake_case ) lowercase_ : int = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) lowercase_ : Optional[Any] = tf.train.load_variable(__snake_case , __snake_case ) lowercase_ : Optional[int] = array # Build TF to PyTorch weights loading map lowercase_ : Any = _build_tf_to_pytorch_map(__snake_case , __snake_case , __snake_case ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue lowercase_ : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase_ : Any = np.transpose(__snake_case , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase_ : Optional[int] = array.squeeze().transpose() else: lowercase_ : Optional[int] = np.transpose(__snake_case , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) lowercase_ : str = torch.from_numpy(__snake_case ) tf_weights.pop(__snake_case , __snake_case ) tf_weights.pop(name + '''/RMSProp''' , __snake_case ) tf_weights.pop(name + '''/RMSProp_1''' , __snake_case ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __snake_case ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def lowercase ( __snake_case : torch.Tensor , __snake_case : nn.Convad ): lowercase_ , lowercase_ : Optional[int] = features.shape[-2:] lowercase_ , lowercase_ : str = conv_layer.stride lowercase_ , lowercase_ : Tuple = conv_layer.kernel_size if in_height % stride_height == 0: lowercase_ : Dict = max(kernel_height - stride_height , 0 ) else: lowercase_ : List[Any] = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase_ : str = max(kernel_width - stride_width , 0 ) else: lowercase_ : int = max(kernel_width - (in_width % stride_width) , 0 ) lowercase_ : int = pad_along_width // 2 lowercase_ : Union[str, Any] = pad_along_width - pad_left lowercase_ : Tuple = pad_along_height // 2 lowercase_ : List[str] = pad_along_height - pad_top lowercase_ : str = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__snake_case , __snake_case , '''constant''' , 0.0 ) class _UpperCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : MobileNetVaConfig , A : int , A : int , A : int , A : Optional[int] = 1 , A : Optional[int] = 1 , A : bool = False , A : Optional[bool] = True , A : Optional[bool or str] = True , ) -> None: super().__init__() lowercase_ : int = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) lowercase_ : Tuple = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase_ : int = nn.Convad( in_channels=A , out_channels=A , kernel_size=A , stride=A , padding=A , groups=A , bias=A , padding_mode='''zeros''' , ) if use_normalization: lowercase_ : Optional[Any] = nn.BatchNormad( num_features=A , eps=config.layer_norm_eps , momentum=0.9997 , affine=A , track_running_stats=A , ) else: lowercase_ : Union[str, Any] = None if use_activation: if isinstance(A , A ): lowercase_ : str = ACTaFN[use_activation] elif isinstance(config.hidden_act , A ): lowercase_ : Any = ACTaFN[config.hidden_act] else: lowercase_ : Tuple = config.hidden_act else: lowercase_ : Tuple = None def A ( self : str , A : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: lowercase_ : List[Any] = apply_tf_padding(A , self.convolution ) lowercase_ : Optional[int] = self.convolution(A ) if self.normalization is not None: lowercase_ : Union[str, Any] = self.normalization(A ) if self.activation is not None: lowercase_ : Optional[int] = self.activation(A ) return features class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = MobileNetVaConfig SCREAMING_SNAKE_CASE_ : int = load_tf_weights_in_mobilenet_va SCREAMING_SNAKE_CASE_ : Optional[Any] = "mobilenet_v1" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any , A : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __A : Union[str, Any] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : str , A : MobileNetVaConfig , A : bool = True ) -> int: super().__init__(A ) lowercase_ : Union[str, Any] = config lowercase_ : List[str] = 32 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase_ : Union[str, Any] = MobileNetVaConvLayer( A , in_channels=config.num_channels , out_channels=A , kernel_size=3 , stride=2 , ) lowercase_ : Optional[Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase_ : List[Any] = nn.ModuleList() for i in range(13 ): lowercase_ : Dict = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=3 , stride=strides[i] , groups=A , ) ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=1 , ) ) lowercase_ : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def A ( self : Any , A : Optional[Any] ) -> Optional[int]: raise NotImplementedError @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : List[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase_ : List[str] = self.conv_stem(A ) lowercase_ : Dict = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase_ : Optional[int] = layer_module(A ) if output_hidden_states: lowercase_ : str = all_hidden_states + (hidden_states,) lowercase_ : Tuple = hidden_states if self.pooler is not None: lowercase_ : Dict = torch.flatten(self.pooler(A ) , start_dim=1 ) else: lowercase_ : Optional[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=A , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : List[str] , A : MobileNetVaConfig ) -> None: super().__init__(A ) lowercase_ : int = config.num_labels lowercase_ : List[str] = MobileNetVaModel(A ) lowercase_ : Union[str, Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase_ : Tuple = nn.Dropout(config.classifier_dropout_prob , inplace=A ) lowercase_ : int = nn.Linear(A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[Any] = self.mobilenet_va(A , output_hidden_states=A , return_dict=A ) lowercase_ : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Dict = self.classifier(self.dropout(A ) ) lowercase_ : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : Optional[Any] = '''single_label_classification''' else: lowercase_ : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : str = MSELoss() if self.num_labels == 1: lowercase_ : List[str] = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(A , A ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[Any] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : str = BCEWithLogitsLoss() lowercase_ : List[Any] = loss_fct(A , A ) if not return_dict: lowercase_ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A , logits=A , hidden_states=outputs.hidden_states , )
33
1
"""simple docstring""" def lowercase ( __snake_case : int , __snake_case : list[int] , __snake_case : int ): def count_of_possible_combinations(__snake_case : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(__snake_case ) def lowercase ( __snake_case : int , __snake_case : list[int] , __snake_case : int ): def count_of_possible_combinations_with_dp_array( __snake_case : int , __snake_case : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowercase_ : Optional[int] = sum( count_of_possible_combinations_with_dp_array(target - item , __snake_case ) for item in array ) lowercase_ : Union[str, Any] = answer return answer lowercase_ : Dict = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(__snake_case , __snake_case ) def lowercase ( __snake_case : int , __snake_case : list[int] , __snake_case : int ): lowercase_ : List[Any] = [0] * (target + 1) lowercase_ : Optional[Any] = 1 for i in range(1 , target + 1 ): for j in range(__snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __A : int = 3 __A : Union[str, Any] = 5 __A : List[str] = [1, 2, 5] print(combination_sum_iv(n, array, target))
33
"""simple docstring""" def lowercase ( __snake_case : list[int] ): lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): for j in range(i + 1 , __snake_case ): if numbers[j] < numbers[i]: lowercase_ , lowercase_ : Optional[int] = numbers[j], numbers[i] return numbers if __name__ == "__main__": __A : int = input('''Enter numbers separated by a comma:\n''').strip() __A : Any = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
33
1
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def lowercase ( __snake_case : Dict , __snake_case : str , __snake_case : Dict=None , __snake_case : Dict=None ): if attention_mask is None: lowercase_ : List[Any] = tf.cast(tf.math.not_equal(__snake_case , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : Optional[int] = OPTConfig SCREAMING_SNAKE_CASE_ : List[str] = {} SCREAMING_SNAKE_CASE_ : Dict = "gelu" def __init__( self : Any , A : List[str] , A : Dict=13 , A : Optional[Any]=7 , A : Dict=True , A : Union[str, Any]=False , A : List[Any]=99 , A : List[str]=16 , A : str=2 , A : List[Any]=4 , A : Union[str, Any]=4 , A : Optional[int]="gelu" , A : Optional[int]=0.1 , A : List[str]=0.1 , A : Optional[Any]=20 , A : Optional[Any]=2 , A : str=1 , A : Union[str, Any]=0 , A : Optional[Any]=16 , A : List[Any]=16 , ) -> Union[str, Any]: lowercase_ : Dict = parent lowercase_ : Dict = batch_size lowercase_ : str = seq_length lowercase_ : Union[str, Any] = is_training lowercase_ : Union[str, Any] = use_labels lowercase_ : List[str] = vocab_size lowercase_ : Union[str, Any] = hidden_size lowercase_ : Optional[int] = num_hidden_layers lowercase_ : List[str] = num_attention_heads lowercase_ : str = intermediate_size lowercase_ : Dict = hidden_act lowercase_ : str = hidden_dropout_prob lowercase_ : Dict = attention_probs_dropout_prob lowercase_ : Dict = max_position_embeddings lowercase_ : int = eos_token_id lowercase_ : Any = pad_token_id lowercase_ : Optional[Any] = bos_token_id lowercase_ : Tuple = embed_dim lowercase_ : Optional[Any] = word_embed_proj_dim lowercase_ : Tuple = False def A ( self : Any ) -> List[str]: lowercase_ : int = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowercase_ : Tuple = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowercase_ : int = tf.concat([input_ids, eos_tensor] , axis=1 ) lowercase_ : int = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=A , **self.config_updates , ) lowercase_ : Tuple = prepare_opt_inputs_dict(A , A ) return config, inputs_dict def A ( self : int , A : Union[str, Any] , A : Tuple ) -> Optional[Any]: lowercase_ : Optional[Any] = TFOPTModel(config=A ) lowercase_ : int = inputs_dict['''input_ids'''] lowercase_ : Dict = input_ids[:1, :] lowercase_ : Tuple = inputs_dict['''attention_mask'''][:1, :] lowercase_ : Optional[int] = 1 # first forward pass lowercase_ : Optional[Any] = model(A , attention_mask=A , use_cache=A ) lowercase_ , lowercase_ : Optional[int] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase_ : List[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase_ : Union[str, Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowercase_ : str = tf.concat([input_ids, next_tokens] , axis=-1 ) lowercase_ : Tuple = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowercase_ : List[str] = model(A , attention_mask=A )[0] lowercase_ : Any = model(A , attention_mask=A , past_key_values=A )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowercase_ : Dict = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowercase_ : Optional[Any] = output_from_no_past[:, -3:, random_slice_idx] lowercase_ : Dict = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A , A , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : int = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = (TFOPTForCausalLM,) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : Dict = ( {"feature-extraction": TFOPTModel, "text-generation": TFOPTForCausalLM} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = False SCREAMING_SNAKE_CASE_ : List[str] = False SCREAMING_SNAKE_CASE_ : List[str] = False SCREAMING_SNAKE_CASE_ : Dict = 10 def A ( self : Union[str, Any] ) -> str: lowercase_ : Dict = TFOPTModelTester(self ) lowercase_ : Dict = ConfigTester(self , config_class=A ) def A ( self : List[str] ) -> str: self.config_tester.run_common_tests() def A ( self : Union[str, Any] ) -> str: lowercase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A ) def A ( self : str ) -> Optional[Any]: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(A : List[str] , A : Any ): if hasattr(A , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(A , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings lowercase_ : str = model_class(config=A ) lowercase_ : Any = _get_word_embedding_weight(A , model.get_input_embeddings() ) lowercase_ : List[Any] = _get_word_embedding_weight(A , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(A ) lowercase_ : Optional[Any] = _get_word_embedding_weight(A , model.get_input_embeddings() ) lowercase_ : Optional[int] = _get_word_embedding_weight(A , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. lowercase_ : List[Any] = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , A ) # check that weights remain the same after resizing lowercase_ : Optional[int] = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: lowercase_ : Any = False self.assertTrue(A ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , A ) lowercase_ : Any = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: lowercase_ : Dict = False self.assertTrue(A ) def lowercase ( __snake_case : int ): return tf.constant(__snake_case , dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = 99 def A ( self : Union[str, Any] ) -> str: lowercase_ : Union[str, Any] = tf.ones((4, 1) , dtype=tf.intaa ) * 2 lowercase_ : Optional[Any] = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) lowercase_ : Dict = input_ids.shape[0] lowercase_ : Optional[int] = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : Any ) -> Union[str, Any]: lowercase_ : int = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) lowercase_ : List[str] = _long_tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) lowercase_ : Dict = tf.not_equal(A , model.config.pad_token_id ) with tf.GradientTape(): lowercase_ : str = model(input_ids=A , attention_mask=A ).last_hidden_state lowercase_ : Union[str, Any] = (1, 11, 5_12) self.assertEqual(output.shape , A ) lowercase_ : Union[str, Any] = tf.constant( [[-0.2873, -1.9218, -0.3033], [-1.2710, -0.1338, -0.1902], [0.4095, 0.1214, -1.3121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , A , atol=4e-3 ) ) lowercase_ : Dict = tf.function(A , jit_compile=A ) lowercase_ : Union[str, Any] = xla_generate(A , A )[0] self.assertTrue(np.allclose(output[:, :3, :3] , A , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): def A ( self : List[str] ) -> Any: super().setUp() lowercase_ : Tuple = '''facebook/opt-350m''' def A ( self : Any ) -> Union[str, Any]: lowercase_ : Optional[int] = TFOPTForCausalLM.from_pretrained(self.path_model ) lowercase_ : str = GPTaTokenizer.from_pretrained(self.path_model ) lowercase_ : Any = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False lowercase_ : str = tokenizer(A , return_tensors='''tf''' , padding=A , add_special_tokens=A ) lowercase_ : str = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) lowercase_ : Any = tf.constant( [ [1.3851, -13.8923, -10.5229, -10.7533, -0.2309, -10.2384, -0.5365, -9.0947, -5.1670], [-4.7073, -10.6276, -3.9415, -21.5242, -0.2822, -0.2822, -0.2822, -0.2822, -0.2822], [0.6247, -3.4229, -8.9179, -1.4297, -14.1650, 1.4146, -9.0218, -0.2703, -0.2703], [6.4783, -1.9913, -10.7926, -2.3336, 1.5092, -0.9974, -6.8213, 1.3477, 1.3477], ] ) self.assertTrue(np.allclose(A , A , atol=1e-4 ) ) lowercase_ : str = tf.function(A , jit_compile=A ) lowercase_ : Dict = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(A , A , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): @property def A ( self : Optional[int] ) -> Any: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def A ( self : Tuple ) -> Optional[Any]: lowercase_ : str = '''facebook/opt-125m''' lowercase_ : int = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] lowercase_ : Tuple = [] lowercase_ : Optional[int] = GPTaTokenizer.from_pretrained(A ) lowercase_ : Any = TFOPTForCausalLM.from_pretrained(A ) for prompt in self.prompts: lowercase_ : int = tokenizer(A , return_tensors='''tf''' ).input_ids lowercase_ : Any = model.generate(A , max_length=10 ) lowercase_ : List[Any] = tokenizer.batch_decode(A , skip_special_tokens=A ) predicted_outputs += generated_string self.assertListEqual(A , A ) def A ( self : Optional[Any] ) -> Union[str, Any]: lowercase_ : int = '''facebook/opt-350m''' lowercase_ : List[str] = GPTaTokenizer.from_pretrained(A ) lowercase_ : Tuple = TFOPTForCausalLM.from_pretrained(A ) lowercase_ : Any = '''left''' # use different length sentences to test batching lowercase_ : Union[str, Any] = [ '''Hello, my dog is a little''', '''Today, I''', ] lowercase_ : List[str] = tokenizer(A , return_tensors='''tf''' , padding=A ) lowercase_ : List[Any] = inputs['''input_ids'''] lowercase_ : List[Any] = model.generate(input_ids=A , attention_mask=inputs['''attention_mask'''] ) lowercase_ : str = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids lowercase_ : Dict = model.generate(input_ids=A ) lowercase_ : Any = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) lowercase_ : Dict = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids lowercase_ : List[str] = model.generate(input_ids=A , max_length=model.config.max_length - num_paddings ) lowercase_ : Optional[int] = tokenizer.batch_decode(A , skip_special_tokens=A ) lowercase_ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=A ) lowercase_ : Tuple = tokenizer.decode(output_padded[0] , skip_special_tokens=A ) lowercase_ : Tuple = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(A , A ) self.assertListEqual(A , [non_padded_sentence, padded_sentence] ) def A ( self : List[str] ) -> Optional[Any]: lowercase_ : Any = '''facebook/opt-350m''' lowercase_ : Optional[Any] = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] lowercase_ : List[Any] = [] lowercase_ : Dict = GPTaTokenizer.from_pretrained(A ) lowercase_ : Tuple = TFOPTForCausalLM.from_pretrained(A ) for prompt in self.prompts: lowercase_ : Optional[Any] = tokenizer(A , return_tensors='''tf''' ).input_ids lowercase_ : Dict = model.generate(A , max_length=10 ) lowercase_ : Optional[Any] = tokenizer.batch_decode(A , skip_special_tokens=A ) predicted_outputs += generated_string self.assertListEqual(A , A )
33
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any] , A : Any , A : Tuple=7 , A : Tuple=3 , A : Optional[Any]=30 , A : List[Any]=4_00 , A : Tuple=True , A : Dict=None , A : List[str]=True , A : Optional[int]=[0.5, 0.5, 0.5] , A : Tuple=[0.5, 0.5, 0.5] , A : List[str]=True , A : List[Any]=1 / 2_55 , A : Union[str, Any]=True , ) -> Tuple: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase_ : Optional[int] = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} lowercase_ : Optional[int] = parent lowercase_ : str = batch_size lowercase_ : Tuple = num_channels lowercase_ : str = min_resolution lowercase_ : Any = max_resolution lowercase_ : str = do_resize lowercase_ : Any = size lowercase_ : Optional[int] = do_normalize lowercase_ : List[str] = image_mean lowercase_ : Optional[Any] = image_std lowercase_ : int = do_rescale lowercase_ : List[str] = rescale_factor lowercase_ : int = do_pad def A ( self : Any ) -> str: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Optional[Any] , A : int , A : int=False ) -> Tuple: if not batched: lowercase_ : Optional[int] = image_inputs[0] if isinstance(A , Image.Image ): lowercase_ , lowercase_ : int = image.size else: lowercase_ , lowercase_ : Tuple = image.shape[1], image.shape[2] if w < h: lowercase_ : int = int(self.size['''shortest_edge'''] * h / w ) lowercase_ : Optional[Any] = self.size['''shortest_edge'''] elif w > h: lowercase_ : Optional[Any] = self.size['''shortest_edge'''] lowercase_ : Optional[int] = int(self.size['''shortest_edge'''] * w / h ) else: lowercase_ : Any = self.size['''shortest_edge'''] lowercase_ : Any = self.size['''shortest_edge'''] else: lowercase_ : Tuple = [] for image in image_inputs: lowercase_ , lowercase_ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowercase_ : Union[str, Any] = max(A , key=lambda A : item[0] )[0] lowercase_ : Optional[Any] = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = YolosImageProcessor if is_vision_available() else None def A ( self : Optional[int] ) -> Optional[int]: lowercase_ : Optional[Any] = YolosImageProcessingTester(self ) @property def A ( self : str ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Optional[int] ) -> List[str]: lowercase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , '''image_mean''' ) ) self.assertTrue(hasattr(A , '''image_std''' ) ) self.assertTrue(hasattr(A , '''do_normalize''' ) ) self.assertTrue(hasattr(A , '''do_resize''' ) ) self.assertTrue(hasattr(A , '''size''' ) ) def A ( self : Dict ) -> Tuple: lowercase_ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowercase_ : Tuple = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A ) def A ( self : Optional[int] ) -> Tuple: pass def A ( self : Tuple ) -> int: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowercase_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ , lowercase_ : Dict = self.image_processor_tester.get_expected_values(A , batched=A ) lowercase_ : str = image_processing(A , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : str ) -> Any: # Initialize image_processing lowercase_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowercase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : int = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Optional[int] = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[Any] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[int]: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowercase_ : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Any = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[str] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[Any]: # Initialize image_processings lowercase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) lowercase_ : Tuple = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowercase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowercase_ : Union[str, Any] = image_processing_a.pad(A , return_tensors='''pt''' ) lowercase_ : List[Any] = image_processing_a(A , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1e-4 ) ) @slow def A ( self : str ) -> List[Any]: # prepare image and target lowercase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowercase_ : List[Any] = json.loads(f.read() ) lowercase_ : Tuple = {'''image_id''': 3_97_69, '''annotations''': target} # encode them lowercase_ : Union[str, Any] = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) lowercase_ : List[Any] = image_processing(images=A , annotations=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Union[str, Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Union[str, Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : Tuple = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : List[str] = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : Any = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify orig_size lowercase_ : List[str] = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : Optional[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) ) @slow def A ( self : List[Any] ) -> Dict: # prepare image, target and masks_path lowercase_ : Tuple = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowercase_ : str = json.loads(f.read() ) lowercase_ : int = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} lowercase_ : List[Any] = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowercase_ : int = YolosImageProcessor(format='''coco_panoptic''' ) lowercase_ : Any = image_processing(images=A , annotations=A , masks_path=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Optional[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Tuple = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : List[Any] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : str = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : List[str] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify masks lowercase_ : Dict = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A ) # verify orig_size lowercase_ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) )
33
1
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool __A : Union[str, Any] = { '''Acehnese Arabic''': '''ace_Arab''', '''Acehnese Latin''': '''ace_Latn''', '''Mesopotamian Arabic''': '''acm_Arab''', '''Ta\'izzi-Adeni Arabic''': '''acq_Arab''', '''Tunisian Arabic''': '''aeb_Arab''', '''Afrikaans''': '''afr_Latn''', '''South Levantine Arabic''': '''ajp_Arab''', '''Akan''': '''aka_Latn''', '''Amharic''': '''amh_Ethi''', '''North Levantine Arabic''': '''apc_Arab''', '''Modern Standard Arabic''': '''arb_Arab''', '''Modern Standard Arabic Romanized''': '''arb_Latn''', '''Najdi Arabic''': '''ars_Arab''', '''Moroccan Arabic''': '''ary_Arab''', '''Egyptian Arabic''': '''arz_Arab''', '''Assamese''': '''asm_Beng''', '''Asturian''': '''ast_Latn''', '''Awadhi''': '''awa_Deva''', '''Central Aymara''': '''ayr_Latn''', '''South Azerbaijani''': '''azb_Arab''', '''North Azerbaijani''': '''azj_Latn''', '''Bashkir''': '''bak_Cyrl''', '''Bambara''': '''bam_Latn''', '''Balinese''': '''ban_Latn''', '''Belarusian''': '''bel_Cyrl''', '''Bemba''': '''bem_Latn''', '''Bengali''': '''ben_Beng''', '''Bhojpuri''': '''bho_Deva''', '''Banjar Arabic''': '''bjn_Arab''', '''Banjar Latin''': '''bjn_Latn''', '''Standard Tibetan''': '''bod_Tibt''', '''Bosnian''': '''bos_Latn''', '''Buginese''': '''bug_Latn''', '''Bulgarian''': '''bul_Cyrl''', '''Catalan''': '''cat_Latn''', '''Cebuano''': '''ceb_Latn''', '''Czech''': '''ces_Latn''', '''Chokwe''': '''cjk_Latn''', '''Central Kurdish''': '''ckb_Arab''', '''Crimean Tatar''': '''crh_Latn''', '''Welsh''': '''cym_Latn''', '''Danish''': '''dan_Latn''', '''German''': '''deu_Latn''', '''Southwestern Dinka''': '''dik_Latn''', '''Dyula''': '''dyu_Latn''', '''Dzongkha''': '''dzo_Tibt''', '''Greek''': '''ell_Grek''', '''English''': '''eng_Latn''', '''Esperanto''': '''epo_Latn''', '''Estonian''': '''est_Latn''', '''Basque''': '''eus_Latn''', '''Ewe''': '''ewe_Latn''', '''Faroese''': '''fao_Latn''', '''Fijian''': '''fij_Latn''', '''Finnish''': '''fin_Latn''', '''Fon''': '''fon_Latn''', '''French''': '''fra_Latn''', '''Friulian''': '''fur_Latn''', '''Nigerian Fulfulde''': '''fuv_Latn''', '''Scottish Gaelic''': '''gla_Latn''', '''Irish''': '''gle_Latn''', '''Galician''': '''glg_Latn''', '''Guarani''': '''grn_Latn''', '''Gujarati''': '''guj_Gujr''', '''Haitian Creole''': '''hat_Latn''', '''Hausa''': '''hau_Latn''', '''Hebrew''': '''heb_Hebr''', '''Hindi''': '''hin_Deva''', '''Chhattisgarhi''': '''hne_Deva''', '''Croatian''': '''hrv_Latn''', '''Hungarian''': '''hun_Latn''', '''Armenian''': '''hye_Armn''', '''Igbo''': '''ibo_Latn''', '''Ilocano''': '''ilo_Latn''', '''Indonesian''': '''ind_Latn''', '''Icelandic''': '''isl_Latn''', '''Italian''': '''ita_Latn''', '''Javanese''': '''jav_Latn''', '''Japanese''': '''jpn_Jpan''', '''Kabyle''': '''kab_Latn''', '''Jingpho''': '''kac_Latn''', '''Kamba''': '''kam_Latn''', '''Kannada''': '''kan_Knda''', '''Kashmiri Arabic''': '''kas_Arab''', '''Kashmiri Devanagari''': '''kas_Deva''', '''Georgian''': '''kat_Geor''', '''Central Kanuri Arabic''': '''knc_Arab''', '''Central Kanuri Latin''': '''knc_Latn''', '''Kazakh''': '''kaz_Cyrl''', '''Kabiyè''': '''kbp_Latn''', '''Kabuverdianu''': '''kea_Latn''', '''Khmer''': '''khm_Khmr''', '''Kikuyu''': '''kik_Latn''', '''Kinyarwanda''': '''kin_Latn''', '''Kyrgyz''': '''kir_Cyrl''', '''Kimbundu''': '''kmb_Latn''', '''Northern Kurdish''': '''kmr_Latn''', '''Kikongo''': '''kon_Latn''', '''Korean''': '''kor_Hang''', '''Lao''': '''lao_Laoo''', '''Ligurian''': '''lij_Latn''', '''Limburgish''': '''lim_Latn''', '''Lingala''': '''lin_Latn''', '''Lithuanian''': '''lit_Latn''', '''Lombard''': '''lmo_Latn''', '''Latgalian''': '''ltg_Latn''', '''Luxembourgish''': '''ltz_Latn''', '''Luba-Kasai''': '''lua_Latn''', '''Ganda''': '''lug_Latn''', '''Luo''': '''luo_Latn''', '''Mizo''': '''lus_Latn''', '''Standard Latvian''': '''lvs_Latn''', '''Magahi''': '''mag_Deva''', '''Maithili''': '''mai_Deva''', '''Malayalam''': '''mal_Mlym''', '''Marathi''': '''mar_Deva''', '''Minangkabau Arabic ''': '''min_Arab''', '''Minangkabau Latin''': '''min_Latn''', '''Macedonian''': '''mkd_Cyrl''', '''Plateau Malagasy''': '''plt_Latn''', '''Maltese''': '''mlt_Latn''', '''Meitei Bengali''': '''mni_Beng''', '''Halh Mongolian''': '''khk_Cyrl''', '''Mossi''': '''mos_Latn''', '''Maori''': '''mri_Latn''', '''Burmese''': '''mya_Mymr''', '''Dutch''': '''nld_Latn''', '''Norwegian Nynorsk''': '''nno_Latn''', '''Norwegian Bokmål''': '''nob_Latn''', '''Nepali''': '''npi_Deva''', '''Northern Sotho''': '''nso_Latn''', '''Nuer''': '''nus_Latn''', '''Nyanja''': '''nya_Latn''', '''Occitan''': '''oci_Latn''', '''West Central Oromo''': '''gaz_Latn''', '''Odia''': '''ory_Orya''', '''Pangasinan''': '''pag_Latn''', '''Eastern Panjabi''': '''pan_Guru''', '''Papiamento''': '''pap_Latn''', '''Western Persian''': '''pes_Arab''', '''Polish''': '''pol_Latn''', '''Portuguese''': '''por_Latn''', '''Dari''': '''prs_Arab''', '''Southern Pashto''': '''pbt_Arab''', '''Ayacucho Quechua''': '''quy_Latn''', '''Romanian''': '''ron_Latn''', '''Rundi''': '''run_Latn''', '''Russian''': '''rus_Cyrl''', '''Sango''': '''sag_Latn''', '''Sanskrit''': '''san_Deva''', '''Santali''': '''sat_Olck''', '''Sicilian''': '''scn_Latn''', '''Shan''': '''shn_Mymr''', '''Sinhala''': '''sin_Sinh''', '''Slovak''': '''slk_Latn''', '''Slovenian''': '''slv_Latn''', '''Samoan''': '''smo_Latn''', '''Shona''': '''sna_Latn''', '''Sindhi''': '''snd_Arab''', '''Somali''': '''som_Latn''', '''Southern Sotho''': '''sot_Latn''', '''Spanish''': '''spa_Latn''', '''Tosk Albanian''': '''als_Latn''', '''Sardinian''': '''srd_Latn''', '''Serbian''': '''srp_Cyrl''', '''Swati''': '''ssw_Latn''', '''Sundanese''': '''sun_Latn''', '''Swedish''': '''swe_Latn''', '''Swahili''': '''swh_Latn''', '''Silesian''': '''szl_Latn''', '''Tamil''': '''tam_Taml''', '''Tatar''': '''tat_Cyrl''', '''Telugu''': '''tel_Telu''', '''Tajik''': '''tgk_Cyrl''', '''Tagalog''': '''tgl_Latn''', '''Thai''': '''tha_Thai''', '''Tigrinya''': '''tir_Ethi''', '''Tamasheq Latin''': '''taq_Latn''', '''Tamasheq Tifinagh''': '''taq_Tfng''', '''Tok Pisin''': '''tpi_Latn''', '''Tswana''': '''tsn_Latn''', '''Tsonga''': '''tso_Latn''', '''Turkmen''': '''tuk_Latn''', '''Tumbuka''': '''tum_Latn''', '''Turkish''': '''tur_Latn''', '''Twi''': '''twi_Latn''', '''Central Atlas Tamazight''': '''tzm_Tfng''', '''Uyghur''': '''uig_Arab''', '''Ukrainian''': '''ukr_Cyrl''', '''Umbundu''': '''umb_Latn''', '''Urdu''': '''urd_Arab''', '''Northern Uzbek''': '''uzn_Latn''', '''Venetian''': '''vec_Latn''', '''Vietnamese''': '''vie_Latn''', '''Waray''': '''war_Latn''', '''Wolof''': '''wol_Latn''', '''Xhosa''': '''xho_Latn''', '''Eastern Yiddish''': '''ydd_Hebr''', '''Yoruba''': '''yor_Latn''', '''Yue Chinese''': '''yue_Hant''', '''Chinese Simplified''': '''zho_Hans''', '''Chinese Traditional''': '''zho_Hant''', '''Standard Malay''': '''zsm_Latn''', '''Zulu''': '''zul_Latn''', } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Dict = "facebook/nllb-200-distilled-600M" SCREAMING_SNAKE_CASE_ : List[str] = ( "This is a tool that translates text from a language to another. It takes three inputs: `text`, which should " "be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, " "which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in " "plain English, such as 'Romanian', or 'Albanian'. It returns the text translated in `tgt_lang`." ) SCREAMING_SNAKE_CASE_ : List[str] = "translator" SCREAMING_SNAKE_CASE_ : str = AutoTokenizer SCREAMING_SNAKE_CASE_ : Any = AutoModelForSeqaSeqLM SCREAMING_SNAKE_CASE_ : str = LANGUAGE_CODES SCREAMING_SNAKE_CASE_ : str = ["text", "text", "text"] SCREAMING_SNAKE_CASE_ : Any = ["text"] def A ( self : List[Any] , A : str , A : Union[str, Any] , A : List[str] ) -> List[str]: if src_lang not in self.lang_to_code: raise ValueError(F'''{src_lang} is not a supported language.''' ) if tgt_lang not in self.lang_to_code: raise ValueError(F'''{tgt_lang} is not a supported language.''' ) lowercase_ : str = self.lang_to_code[src_lang] lowercase_ : Union[str, Any] = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( A , return_tensors='''pt''' , src_lang=A , tgt_lang=A ) def A ( self : str , A : Optional[int] ) -> List[str]: return self.model.generate(**A ) def A ( self : Any , A : Union[str, Any] ) -> List[str]: return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=A )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0 ): lowercase_ : str = 0 lowercase_ : List[Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
33
1
"""simple docstring""" # Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union __A : List[str] = re.compile(R'''^(?P<major>\d+)''' R'''\.(?P<minor>\d+)''' R'''\.(?P<patch>\d+)$''') @total_ordering @dataclass class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : str SCREAMING_SNAKE_CASE_ : Optional[str] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None SCREAMING_SNAKE_CASE_ : Optional[Union[str, int]] = None def A ( self : Optional[int] ) -> Union[str, Any]: lowercase_ , lowercase_ , lowercase_ : Union[str, Any] = _str_to_version_tuple(self.version_str ) def __repr__( self : int ) -> List[Any]: return F'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def A ( self : Any ) -> Union[str, Any]: return self.major, self.minor, self.patch def A ( self : Dict , A : List[Any] ) -> Tuple: if isinstance(A , A ): return Version(A ) elif isinstance(A , A ): return other raise TypeError(F'''{other} (type {type(A )}) cannot be compared to version.''' ) def __eq__( self : Union[str, Any] , A : str ) -> List[str]: try: lowercase_ : Optional[int] = self._validate_operand(A ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : int , A : str ) -> Union[str, Any]: lowercase_ : Optional[Any] = self._validate_operand(A ) return self.tuple < other.tuple def __hash__( self : Dict ) -> Optional[Any]: return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def A ( cls : int , A : List[Any] ) -> List[str]: lowercase_ : Dict = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def A ( self : Optional[int] ) -> str: return self.version_str def lowercase ( __snake_case : Any ): lowercase_ : int = _VERSION_REG.match(__snake_case ) if not res: raise ValueError(F'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(__snake_case ) for v in [res.group('''major''' ), res.group('''minor''' ), res.group('''patch''' )] ) def lowercase ( __snake_case : int ): return ".".join(str(__snake_case ) for v in version_tuple )
33
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
1
"""simple docstring""" from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def lowercase ( __snake_case : np.ndarray , __snake_case : np.ndarray , __snake_case : np.ndarray , __snake_case : int , __snake_case : int ): lowercase_ : Optional[int] = cva.getAffineTransform(__snake_case , __snake_case ) return cva.warpAffine(__snake_case , __snake_case , (rows, cols) ) if __name__ == "__main__": # read original image __A : Tuple = cva.imread( str(Path(__file__).resolve().parent.parent / '''image_data''' / '''lena.jpg''') ) # turn image in gray scale value __A : int = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape __A , __A : Union[str, Any] = gray_img.shape # set different points to rotate image __A : Optional[Any] = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) __A : int = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) __A : List[Any] = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) __A : str = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list __A : Union[str, Any] = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations __A : Optional[int] = plt.figure(1) __A : int = ['''Original''', '''Rotation 1''', '''Rotation 2''', '''Rotation 3'''] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, '''gray''') plt.title(titles[i]) plt.axis('''off''') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
33
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
1
"""simple docstring""" import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : str = "char" SCREAMING_SNAKE_CASE_ : Any = "bpe" SCREAMING_SNAKE_CASE_ : Optional[int] = "wp" __A : int = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Any = ["image_processor", "char_tokenizer"] SCREAMING_SNAKE_CASE_ : Any = "ViTImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] = "MgpstrTokenizer" def __init__( self : List[str] , A : Union[str, Any]=None , A : str=None , **A : Optional[Any] ) -> str: lowercase_ : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A , ) lowercase_ : List[Any] = kwargs.pop('''feature_extractor''' ) lowercase_ : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) lowercase_ : List[str] = tokenizer lowercase_ : Union[str, Any] = AutoTokenizer.from_pretrained('''gpt2''' ) lowercase_ : List[Any] = AutoTokenizer.from_pretrained('''bert-base-uncased''' ) super().__init__(A , A ) def __call__( self : List[Any] , A : int=None , A : Tuple=None , A : List[Any]=None , **A : Union[str, Any] ) -> int: if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: lowercase_ : Optional[Any] = self.image_processor(A , return_tensors=A , **A ) if text is not None: lowercase_ : Union[str, Any] = self.char_tokenizer(A , return_tensors=A , **A ) if text is None: return inputs elif images is None: return encodings else: lowercase_ : List[str] = encodings['''input_ids'''] return inputs def A ( self : Optional[Any] , A : int ) -> Optional[int]: lowercase_ , lowercase_ , lowercase_ : Any = sequences lowercase_ : Tuple = char_preds.size(0 ) lowercase_ , lowercase_ : Optional[int] = self._decode_helper(A , '''char''' ) lowercase_ , lowercase_ : Dict = self._decode_helper(A , '''bpe''' ) lowercase_ , lowercase_ : Optional[int] = self._decode_helper(A , '''wp''' ) lowercase_ : int = [] lowercase_ : Optional[int] = [] for i in range(A ): lowercase_ : Dict = [char_scores[i], bpe_scores[i], wp_scores[i]] lowercase_ : List[Any] = [char_strs[i], bpe_strs[i], wp_strs[i]] lowercase_ : Dict = scores.index(max(A ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) lowercase_ : List[Any] = {} lowercase_ : Any = final_strs lowercase_ : List[Any] = final_scores lowercase_ : Optional[Any] = char_strs lowercase_ : Union[str, Any] = bpe_strs lowercase_ : Dict = wp_strs return out def A ( self : Dict , A : List[Any] , A : Dict ) -> Optional[Any]: if format == DecodeType.CHARACTER: lowercase_ : List[str] = self.char_decode lowercase_ : str = 1 lowercase_ : str = '''[s]''' elif format == DecodeType.BPE: lowercase_ : Optional[int] = self.bpe_decode lowercase_ : List[Any] = 2 lowercase_ : Any = '''#''' elif format == DecodeType.WORDPIECE: lowercase_ : Optional[Any] = self.wp_decode lowercase_ : Optional[int] = 1_02 lowercase_ : str = '''[SEP]''' else: raise ValueError(F'''Format {format} is not supported.''' ) lowercase_ , lowercase_ : Union[str, Any] = [], [] lowercase_ : Dict = pred_logits.size(0 ) lowercase_ : Dict = pred_logits.size(1 ) lowercase_ , lowercase_ : Dict = pred_logits.topk(1 , dim=-1 , largest=A , sorted=A ) lowercase_ : List[str] = preds_index.view(-1 , A )[:, 1:] lowercase_ : Dict = decoder(A ) lowercase_ , lowercase_ : Dict = torch.nn.functional.softmax(A , dim=2 ).max(dim=2 ) lowercase_ : Optional[int] = preds_max_prob[:, 1:] for index in range(A ): lowercase_ : Union[str, Any] = preds_str[index].find(A ) lowercase_ : Union[str, Any] = preds_str[index][:pred_eos] lowercase_ : Dict = preds_index[index].cpu().tolist() lowercase_ : int = pred_index.index(A ) if eos_token in pred_index else -1 lowercase_ : List[str] = preds_max_prob[index][: pred_eos_index + 1] lowercase_ : Optional[int] = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(A ) conf_scores.append(A ) return dec_strs, conf_scores def A ( self : str , A : Optional[int] ) -> Any: lowercase_ : Tuple = [seq.replace(''' ''' , '''''' ) for seq in self.char_tokenizer.batch_decode(A )] return decode_strs def A ( self : Tuple , A : Union[str, Any] ) -> Optional[int]: return self.bpe_tokenizer.batch_decode(A ) def A ( self : Dict , A : int ) -> str: lowercase_ : str = [seq.replace(''' ''' , '''''' ) for seq in self.wp_tokenizer.batch_decode(A )] return decode_strs
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0_0 ): lowercase_ , lowercase_ : str = 1, 1 lowercase_ : List[str] = 2 while True: lowercase_ : Tuple = 0 lowercase_ : List[Any] = fa + fa lowercase_ , lowercase_ : Optional[int] = fa, f index += 1 for _ in str(__snake_case ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList __A : Optional[int] = ['''\nclass''', '''\ndef''', '''\n#''', '''\n@''', '''\nprint''', '''\nif'''] class _UpperCAmelCase ( _A ): def __init__( self : int , A : Optional[int] , A : Any , A : Optional[int]=None , A : Optional[int]=1 ) -> Union[str, Any]: lowercase_ : List[Any] = tokenizer lowercase_ : Any = dataset lowercase_ : Dict = len(A ) if n_tasks is None else n_tasks lowercase_ : Tuple = n_copies def __iter__( self : Optional[int] ) -> Union[str, Any]: lowercase_ : Optional[Any] = [] for task in range(self.n_tasks ): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]['''prompt'''].strip() ) lowercase_ : str = self.tokenizer(A , padding=A , return_tensors='''pt''' ) for task in range(self.n_tasks ): for _ in range(self.n_copies ): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class _UpperCAmelCase ( _A ): def __init__( self : Dict , A : Optional[Any] , A : Tuple , A : Dict ) -> List[Any]: lowercase_ : Any = start_length lowercase_ : str = eof_strings lowercase_ : Optional[Any] = tokenizer def __call__( self : Optional[Any] , A : Tuple , A : int , **A : Optional[Any] ) -> Optional[int]: lowercase_ : Optional[Any] = self.tokenizer.batch_decode(input_ids[:, self.start_length :] ) lowercase_ : Dict = [] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings ) ) return all(A ) def lowercase ( __snake_case : Optional[int] ): lowercase_ : Any = re.split('''(%s)''' % '''|'''.join(__snake_case ) , __snake_case ) # last string should be "" return "".join(string_list[:-2] ) def lowercase ( __snake_case : Any , __snake_case : str , __snake_case : List[str] , __snake_case : Any , __snake_case : Dict , __snake_case : Dict=2_0 , **__snake_case : Union[str, Any] ): lowercase_ : List[Any] = defaultdict(__snake_case ) # dict of list of generated tokens for step, batch in tqdm(enumerate(__snake_case ) ): with torch.no_grad(): lowercase_ : Tuple = batch['''ids'''].shape[-1] lowercase_ : List[str] = accelerator.unwrap_model(__snake_case ).generate( input_ids=batch['''ids'''][:, : batch['''input_len''']] , num_return_sequences=__snake_case , **__snake_case ) # each task is generated batch_size times lowercase_ : Union[str, Any] = batch['''task_id'''].repeat(__snake_case ) lowercase_ : int = accelerator.pad_across_processes( __snake_case , dim=1 , pad_index=tokenizer.pad_token_id ) lowercase_ , lowercase_ : str = accelerator.gather((generated_tokens, generated_tasks) ) lowercase_ : Tuple = generated_tokens.cpu().numpy() lowercase_ : Any = generated_tasks.cpu().numpy() for task, generated_tokens in zip(__snake_case , __snake_case ): gen_token_dict[task].append(__snake_case ) lowercase_ : Union[str, Any] = [[] for _ in range(__snake_case )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: lowercase_ : Any = tokenizer.decode(__snake_case , skip_special_tokens=__snake_case , clean_up_tokenization_spaces=__snake_case ) code_gens[task].append(remove_last_block(__snake_case ) ) return code_gens def lowercase ( ): # Setup configuration lowercase_ : List[str] = HfArgumentParser(__snake_case ) lowercase_ : Tuple = parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric lowercase_ : int = args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing lowercase_ : Tuple = '''false''' if args.num_workers is None: lowercase_ : str = multiprocessing.cpu_count() # Use dataset load to feed to accelerate lowercase_ : Any = Accelerator() set_seed(args.seed , device_specific=__snake_case ) # Load model and tokenizer lowercase_ : Any = AutoTokenizer.from_pretrained(args.model_ckpt ) lowercase_ : Union[str, Any] = tokenizer.eos_token lowercase_ : Dict = AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings lowercase_ : List[Any] = { '''do_sample''': args.do_sample, '''temperature''': args.temperature, '''max_new_tokens''': args.max_new_tokens, '''top_p''': args.top_p, '''top_k''': args.top_k, '''stopping_criteria''': StoppingCriteriaList([EndOfFunctionCriteria(0 , __snake_case , __snake_case )] ), } # Load evaluation dataset and metric lowercase_ : Optional[Any] = load_dataset('''openai_humaneval''' ) lowercase_ : Any = load_metric('''code_eval''' ) lowercase_ : str = args.num_tasks if args.num_tasks is not None else len(human_eval['''test'''] ) lowercase_ : Tuple = args.n_samples // args.batch_size lowercase_ : Dict = TokenizedDataset(__snake_case , human_eval['''test'''] , n_copies=__snake_case , n_tasks=__snake_case ) # do not confuse args.batch_size, which is actually the num_return_sequences lowercase_ : Optional[int] = DataLoader(__snake_case , batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: lowercase_ : Union[str, Any] = code_eval_metric.compute(references=[''''''] , predictions=[['''''']] ) except ValueError as exception: print( '''Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL="1"`''' ''' flag to enable code evaluation.''' ) raise exception lowercase_ , lowercase_ : Dict = accelerator.prepare(__snake_case , __snake_case ) lowercase_ : Optional[int] = complete_code( __snake_case , __snake_case , __snake_case , __snake_case , n_tasks=__snake_case , batch_size=args.batch_size , **__snake_case , ) if accelerator.is_main_process: lowercase_ : Union[str, Any] = [] for task in tqdm(range(__snake_case ) ): lowercase_ : str = human_eval['''test'''][task]['''test'''] lowercase_ : List[Any] = F'''check({human_eval['test'][task]['entry_point']})''' references.append('''\n''' + test_func + '''\n''' + entry_point ) # Evaluate completions with "code_eval" metric lowercase_ , lowercase_ : Optional[int] = code_eval_metric.compute( references=__snake_case , predictions=__snake_case , num_workers=args.num_workers ) print(F'''Results: {pass_at_k}''' ) # Save results to json file with open(args.output_file , '''w''' ) as fp: json.dump(__snake_case , __snake_case ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : Union[str, Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = "vit_mae" def __init__( self : Dict , A : List[str]=7_68 , A : Any=12 , A : Union[str, Any]=12 , A : Tuple=30_72 , A : Any="gelu" , A : Tuple=0.0 , A : List[str]=0.0 , A : Tuple=0.02 , A : Tuple=1e-12 , A : int=2_24 , A : Dict=16 , A : int=3 , A : Tuple=True , A : Tuple=16 , A : Optional[Any]=5_12 , A : Union[str, Any]=8 , A : List[Any]=20_48 , A : Dict=0.75 , A : Any=False , **A : Optional[int] , ) -> Union[str, Any]: super().__init__(**A ) lowercase_ : List[Any] = hidden_size lowercase_ : str = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : List[Any] = hidden_dropout_prob lowercase_ : int = attention_probs_dropout_prob lowercase_ : int = initializer_range lowercase_ : Dict = layer_norm_eps lowercase_ : Optional[Any] = image_size lowercase_ : str = patch_size lowercase_ : Dict = num_channels lowercase_ : Any = qkv_bias lowercase_ : Union[str, Any] = decoder_num_attention_heads lowercase_ : Optional[Any] = decoder_hidden_size lowercase_ : List[str] = decoder_num_hidden_layers lowercase_ : List[Any] = decoder_intermediate_size lowercase_ : Optional[Any] = mask_ratio lowercase_ : Optional[Any] = norm_pix_loss
33
1
"""simple docstring""" from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class _UpperCAmelCase : def __init__( self : Union[str, Any] , A : str , A : List[Any]=13 , A : int=7 , A : str=True , A : Optional[int]=True , A : Any=True , A : List[str]=True , A : Tuple=99 , A : Union[str, Any]=32 , A : Optional[Any]=2 , A : Dict=4 , A : int=37 , A : List[Any]="gelu" , A : Any=0.1 , A : Dict=0.1 , A : List[str]=5_12 , A : Any=16 , A : str=2 , A : List[Any]=0.02 , A : int=3 , A : List[str]=4 , A : Optional[int]=None , ) -> Union[str, Any]: lowercase_ : List[str] = parent lowercase_ : Tuple = 13 lowercase_ : Any = 7 lowercase_ : List[Any] = True lowercase_ : Union[str, Any] = True lowercase_ : Dict = True lowercase_ : str = True lowercase_ : str = 99 lowercase_ : Union[str, Any] = 3_84 lowercase_ : List[str] = 2 lowercase_ : Optional[int] = 4 lowercase_ : Dict = 37 lowercase_ : int = '''gelu''' lowercase_ : List[Any] = 0.1 lowercase_ : List[str] = 0.1 lowercase_ : int = 5_12 lowercase_ : Optional[Any] = 16 lowercase_ : Optional[int] = 2 lowercase_ : Optional[int] = 0.02 lowercase_ : int = 3 lowercase_ : Union[str, Any] = 4 lowercase_ : Any = 1_28 lowercase_ : Union[str, Any] = 2 lowercase_ : int = 9 lowercase_ : Dict = 1 lowercase_ : str = None def A ( self : Any ) -> List[Any]: lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : Dict = None if self.use_input_mask: lowercase_ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : Optional[Any] = None if self.use_token_type_ids: lowercase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase_ : Union[str, Any] = None lowercase_ : Any = None lowercase_ : str = None if self.use_labels: lowercase_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ : Optional[int] = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[Any] , A : Union[str, Any] , A : Tuple , A : Optional[int] , A : Tuple , A : Any , A : Optional[Any] , A : str ) -> str: lowercase_ : Optional[Any] = TFConvBertModel(config=A ) lowercase_ : Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : int = [input_ids, input_mask] lowercase_ : Optional[Any] = model(A ) lowercase_ : Union[str, Any] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Union[str, Any] , A : Optional[int] , A : str , A : List[Any] , A : int , A : Any , A : Tuple , A : Tuple ) -> str: lowercase_ : Any = TFConvBertForMaskedLM(config=A ) lowercase_ : Dict = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : List[Any] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : int , A : Dict , A : Any , A : List[str] , A : List[Any] , A : Optional[int] , A : Optional[Any] , A : Tuple ) -> Optional[int]: lowercase_ : str = self.num_labels lowercase_ : List[Any] = TFConvBertForSequenceClassification(config=A ) lowercase_ : List[Any] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : str = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : List[Any] , A : List[str] , A : Tuple , A : int , A : Optional[Any] , A : Any , A : str , A : int ) -> Union[str, Any]: lowercase_ : Any = self.num_choices lowercase_ : List[Any] = TFConvBertForMultipleChoice(config=A ) lowercase_ : List[str] = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Optional[Any] = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Optional[Any] = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Dict = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Union[str, Any] , A : List[Any] , A : Dict , A : str , A : Optional[Any] , A : List[Any] , A : Union[str, Any] , A : Optional[Any] ) -> int: lowercase_ : Dict = self.num_labels lowercase_ : List[str] = TFConvBertForTokenClassification(config=A ) lowercase_ : List[Any] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : Optional[Any] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Tuple , A : int , A : Optional[Any] , A : Any , A : str , A : Union[str, Any] , A : int , A : Optional[int] ) -> List[Any]: lowercase_ : Any = TFConvBertForQuestionAnswering(config=A ) lowercase_ : Optional[int] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : Dict = model(A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : List[Any] ) -> Optional[Any]: lowercase_ : Optional[int] = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Optional[Any] = config_and_inputs lowercase_ : List[Any] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : Any = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[Any] = False SCREAMING_SNAKE_CASE_ : Optional[Any] = False SCREAMING_SNAKE_CASE_ : List[Any] = False def A ( self : Optional[Any] ) -> Tuple: lowercase_ : Union[str, Any] = TFConvBertModelTester(self ) lowercase_ : Dict = ConfigTester(self , config_class=A , hidden_size=37 ) def A ( self : Tuple ) -> List[str]: self.config_tester.run_common_tests() def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : Optional[Any] ) -> Tuple: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A ) def A ( self : Tuple ) -> Any: lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A ) def A ( self : Dict ) -> Tuple: lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) def A ( self : Optional[int] ) -> str: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A ) def A ( self : str ) -> int: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) @slow def A ( self : Optional[Any] ) -> Dict: lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Optional[Any] = True lowercase_ : str = True if hasattr(A , '''use_cache''' ): lowercase_ : int = True lowercase_ : Dict = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length ) lowercase_ : Union[str, Any] = getattr(self.model_tester , '''key_length''' , A ) for model_class in self.all_model_classes: lowercase_ : List[str] = self._prepare_for_class(A , A ) lowercase_ : str = model_class(A ) lowercase_ : Any = len(model(A ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A , saved_model=A ) lowercase_ : Optional[int] = os.path.join(A , '''saved_model''' , '''1''' ) lowercase_ : List[Any] = tf.keras.models.load_model(A ) lowercase_ : List[str] = model(A ) if self.is_encoder_decoder: lowercase_ : Any = outputs['''encoder_hidden_states'''] lowercase_ : List[str] = outputs['''encoder_attentions'''] else: lowercase_ : Optional[int] = outputs['''hidden_states'''] lowercase_ : Tuple = outputs['''attentions'''] self.assertEqual(len(A ) , A ) lowercase_ : Any = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A ) , A ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def A ( self : Optional[Any] ) -> Tuple: lowercase_ : List[Any] = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''' ) self.assertIsNotNone(A ) def A ( self : List[str] ) -> Optional[int]: lowercase_ , lowercase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : List[str] = True lowercase_ : Optional[int] = getattr(self.model_tester , '''decoder_seq_length''' , self.model_tester.seq_length ) lowercase_ : List[str] = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length ) lowercase_ : Any = getattr(self.model_tester , '''key_length''' , A ) lowercase_ : List[Any] = getattr(self.model_tester , '''key_length''' , A ) def check_decoder_attentions_output(A : str ): lowercase_ : int = len(A ) self.assertEqual(out_len % 2 , 0 ) lowercase_ : Dict = outputs.decoder_attentions self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A : Any ): lowercase_ : int = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: lowercase_ : Optional[Any] = True lowercase_ : List[Any] = False lowercase_ : int = model_class(A ) lowercase_ : Union[str, Any] = model(self._prepare_for_class(A , A ) ) lowercase_ : str = len(A ) self.assertEqual(config.output_hidden_states , A ) check_encoder_attentions_output(A ) if self.is_encoder_decoder: lowercase_ : Any = model_class(A ) lowercase_ : List[str] = model(self._prepare_for_class(A , A ) ) self.assertEqual(config.output_hidden_states , A ) check_decoder_attentions_output(A ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowercase_ : Dict = True lowercase_ : Any = model_class(A ) lowercase_ : Tuple = model(self._prepare_for_class(A , A ) ) self.assertEqual(config.output_hidden_states , A ) check_encoder_attentions_output(A ) # Check attention is always last and order is fine lowercase_ : Optional[Any] = True lowercase_ : Dict = True lowercase_ : Any = model_class(A ) lowercase_ : Optional[Any] = model(self._prepare_for_class(A , A ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A ) ) self.assertEqual(model.config.output_hidden_states , A ) check_encoder_attentions_output(A ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : List[str] ) -> Dict: lowercase_ : Dict = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''' ) lowercase_ : Union[str, Any] = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase_ : Union[str, Any] = model(A )[0] lowercase_ : List[Any] = [1, 6, 7_68] self.assertEqual(output.shape , A ) lowercase_ : Union[str, Any] = tf.constant( [ [ [-0.03475493, -0.4686034, -0.30638832], [0.22637248, -0.26988646, -0.7423424], [0.10324868, -0.45013508, -0.58280784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A , atol=1e-4 )
33
"""simple docstring""" def lowercase ( __snake_case : int ): if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: lowercase_ : Dict = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase ( __snake_case : int ): lowercase_ : str = 0 lowercase_ : List[str] = 2 while digits < n: index += 1 lowercase_ : Any = len(str(fibonacci(__snake_case ) ) ) return index def lowercase ( __snake_case : int = 1_0_0_0 ): return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" import logging from transformers.configuration_utils import PretrainedConfig __A : Any = logging.getLogger(__name__) class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Tuple = "masked_bert" def __init__( self : Optional[int] , A : int=3_05_22 , A : int=7_68 , A : List[Any]=12 , A : Union[str, Any]=12 , A : List[str]=30_72 , A : Dict="gelu" , A : Any=0.1 , A : int=0.1 , A : Optional[Any]=5_12 , A : Union[str, Any]=2 , A : Any=0.02 , A : str=1e-12 , A : Optional[int]=0 , A : Union[str, Any]="topK" , A : Union[str, Any]="constant" , A : Optional[int]=0.0 , **A : List[str] , ) -> int: super().__init__(pad_token_id=A , **A ) lowercase_ : str = vocab_size lowercase_ : List[str] = hidden_size lowercase_ : List[Any] = num_hidden_layers lowercase_ : int = num_attention_heads lowercase_ : Union[str, Any] = hidden_act lowercase_ : Any = intermediate_size lowercase_ : Tuple = hidden_dropout_prob lowercase_ : Optional[Any] = attention_probs_dropout_prob lowercase_ : Tuple = max_position_embeddings lowercase_ : Union[str, Any] = type_vocab_size lowercase_ : Optional[int] = initializer_range lowercase_ : List[Any] = layer_norm_eps lowercase_ : Any = pruning_method lowercase_ : Dict = mask_init lowercase_ : Optional[Any] = mask_scale
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __A : List[str] = { '''configuration_mobilenet_v2''': [ '''MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileNetV2Config''', '''MobileNetV2OnnxConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''MobileNetV2FeatureExtractor'''] __A : Optional[int] = ['''MobileNetV2ImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileNetV2ForImageClassification''', '''MobileNetV2ForSemanticSegmentation''', '''MobileNetV2Model''', '''MobileNetV2PreTrainedModel''', '''load_tf_weights_in_mobilenet_v2''', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class _UpperCAmelCase : def __init__( self : List[str] , A : Any , ) -> Dict: lowercase_ : Tuple = parent lowercase_ : str = 13 lowercase_ : Optional[Any] = 7 lowercase_ : Any = True lowercase_ : str = True lowercase_ : List[str] = True lowercase_ : int = True lowercase_ : Dict = True lowercase_ : int = False lowercase_ : Dict = False lowercase_ : Union[str, Any] = False lowercase_ : List[Any] = 2 lowercase_ : Optional[int] = 99 lowercase_ : List[Any] = 0 lowercase_ : Dict = 32 lowercase_ : List[Any] = 2 lowercase_ : Tuple = 4 lowercase_ : List[Any] = 0.1 lowercase_ : List[Any] = 0.1 lowercase_ : Optional[Any] = 5_12 lowercase_ : Optional[Any] = 16 lowercase_ : List[str] = 2 lowercase_ : str = 0.02 lowercase_ : Any = 3 lowercase_ : List[str] = 4 lowercase_ : Dict = '''last''' lowercase_ : int = True lowercase_ : str = None lowercase_ : Dict = 0 def A ( self : List[str] ) -> List[Any]: lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] , dtype=tf.floataa ) lowercase_ : Any = None if self.use_input_lengths: lowercase_ : Optional[Any] = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length lowercase_ : List[Any] = None if self.use_token_type_ids: lowercase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) lowercase_ : Optional[int] = None lowercase_ : str = None lowercase_ : Optional[Any] = None if self.use_labels: lowercase_ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : Optional[int] = ids_tensor([self.batch_size] , 2 , dtype=tf.floataa ) lowercase_ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ : Any = FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , bos_token_id=self.bos_token_id , ) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def A ( self : int , A : Dict , A : Dict , A : str , A : Tuple , A : Optional[Any] , A : str , A : Union[str, Any] , A : Dict , A : Optional[int] , ) -> int: lowercase_ : str = TFFlaubertModel(config=A ) lowercase_ : int = {'''input_ids''': input_ids, '''lengths''': input_lengths, '''langs''': token_type_ids} lowercase_ : int = model(A ) lowercase_ : Any = [input_ids, input_mask] lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[Any] , A : int , A : List[str] , A : Dict , A : int , A : int , A : Dict , A : Optional[int] , A : Dict , A : int , ) -> Optional[Any]: lowercase_ : int = TFFlaubertWithLMHeadModel(A ) lowercase_ : List[Any] = {'''input_ids''': input_ids, '''lengths''': input_lengths, '''langs''': token_type_ids} lowercase_ : List[str] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[str] , A : Optional[int] , A : Tuple , A : Dict , A : List[Any] , A : Dict , A : Any , A : List[Any] , A : List[Any] , A : str , ) -> Union[str, Any]: lowercase_ : List[Any] = TFFlaubertForQuestionAnsweringSimple(A ) lowercase_ : Tuple = {'''input_ids''': input_ids, '''lengths''': input_lengths} lowercase_ : List[str] = model(A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : int , A : Optional[Any] , A : int , A : List[str] , A : Optional[Any] , A : Tuple , A : Dict , A : Any , A : Any , A : str , ) -> Optional[int]: lowercase_ : str = TFFlaubertForSequenceClassification(A ) lowercase_ : Any = {'''input_ids''': input_ids, '''lengths''': input_lengths} lowercase_ : Union[str, Any] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A ( self : List[Any] , A : Tuple , A : int , A : Dict , A : Any , A : int , A : Optional[int] , A : str , A : str , A : int , ) -> Optional[int]: lowercase_ : Optional[Any] = self.num_labels lowercase_ : List[Any] = TFFlaubertForTokenClassification(config=A ) lowercase_ : int = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : int = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[str] , A : str , A : Any , A : int , A : Dict , A : Tuple , A : List[Any] , A : Optional[Any] , A : List[Any] , A : Optional[int] , ) -> Union[str, Any]: lowercase_ : Union[str, Any] = self.num_choices lowercase_ : Union[str, Any] = TFFlaubertForMultipleChoice(config=A ) lowercase_ : Union[str, Any] = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Any = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Dict = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Any = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } lowercase_ : Any = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Optional[int] ) -> str: lowercase_ : List[Any] = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Dict = config_and_inputs lowercase_ : Optional[int] = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''langs''': token_type_ids, '''lengths''': input_lengths, } return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : Dict = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( { "feature-extraction": TFFlaubertModel, "fill-mask": TFFlaubertWithLMHeadModel, "question-answering": TFFlaubertForQuestionAnsweringSimple, "text-classification": TFFlaubertForSequenceClassification, "token-classification": TFFlaubertForTokenClassification, "zero-shot": TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[Any] = False def A ( self : Any , A : Any , A : Union[str, Any] , A : Optional[int] , A : int , A : str ) -> Any: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('''Fast''' ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = TFFlaubertModelTester(self ) lowercase_ : Optional[Any] = ConfigTester(self , config_class=A , emb_dim=37 ) def A ( self : List[str] ) -> Dict: self.config_tester.run_common_tests() def A ( self : List[str] ) -> int: lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A ) def A ( self : List[str] ) -> int: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A ) def A ( self : List[str] ) -> int: lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A ) def A ( self : Any ) -> List[Any]: lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A ) def A ( self : Optional[int] ) -> str: lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A ) def A ( self : Optional[Any] ) -> Tuple: lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A ) @slow def A ( self : List[Any] ) -> Optional[Any]: for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFFlaubertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : str ) -> int: lowercase_ : Any = TFFlaubertModel.from_pretrained('''jplu/tf-flaubert-small-cased''' ) lowercase_ : Optional[int] = tf.convert_to_tensor( [[0, 1_58, 7_35, 25_92, 14_24, 67_27, 82, 1]] , dtype=tf.intaa , ) # "J'aime flaubert !" lowercase_ : int = model(A )[0] lowercase_ : str = tf.TensorShape((1, 8, 5_12) ) self.assertEqual(output.shape , A ) # compare the actual values for a slice. lowercase_ : int = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
33
"""simple docstring""" from __future__ import annotations __A : List[Any] = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __A : str = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): lowercase_ : float = -1 for j in range(i + 1 , __snake_case ): if arr[i] < arr[j]: lowercase_ : List[str] = arr[j] break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] for i, outer in enumerate(__snake_case ): lowercase_ : float = -1 for inner in arr[i + 1 :]: if outer < inner: lowercase_ : List[Any] = inner break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = len(__snake_case ) lowercase_ : list[float] = [] lowercase_ : list[float] = [-1] * arr_size for index in reversed(range(__snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowercase_ : Optional[Any] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __A : int = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __A : List[str] = { '''configuration_perceiver''': ['''PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PerceiverConfig''', '''PerceiverOnnxConfig'''], '''tokenization_perceiver''': ['''PerceiverTokenizer'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = ['''PerceiverFeatureExtractor'''] __A : List[Any] = ['''PerceiverImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = [ '''PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PerceiverForImageClassificationConvProcessing''', '''PerceiverForImageClassificationFourier''', '''PerceiverForImageClassificationLearned''', '''PerceiverForMaskedLM''', '''PerceiverForMultimodalAutoencoding''', '''PerceiverForOpticalFlow''', '''PerceiverForSequenceClassification''', '''PerceiverLayer''', '''PerceiverModel''', '''PerceiverPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __A : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __A : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __A : Optional[Any] = { '''configuration_mobilevit''': ['''MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileViTConfig''', '''MobileViTOnnxConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = ['''MobileViTFeatureExtractor'''] __A : Optional[Any] = ['''MobileViTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ '''MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileViTForImageClassification''', '''MobileViTForSemanticSegmentation''', '''MobileViTModel''', '''MobileViTPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileViTForImageClassification''', '''TFMobileViTForSemanticSegmentation''', '''TFMobileViTModel''', '''TFMobileViTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys __A : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" def lowercase ( __snake_case : int ): if not isinstance(__snake_case , __snake_case ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
33
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __A : List[str] = logging.get_logger(__name__) __A : List[str] = { '''microsoft/focalnet-tiny''': '''https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json''', } class _UpperCAmelCase ( _A , _A ): SCREAMING_SNAKE_CASE_ : str = "focalnet" def __init__( self : Tuple , A : Tuple=2_24 , A : Tuple=4 , A : Union[str, Any]=3 , A : Optional[Any]=96 , A : List[Any]=False , A : int=[1_92, 3_84, 7_68, 7_68] , A : Tuple=[2, 2, 6, 2] , A : Union[str, Any]=[2, 2, 2, 2] , A : int=[3, 3, 3, 3] , A : Optional[int]="gelu" , A : Optional[Any]=4.0 , A : Optional[Any]=0.0 , A : Any=0.1 , A : Optional[int]=False , A : Any=1e-4 , A : Optional[int]=False , A : List[Any]=False , A : Dict=False , A : Dict=0.02 , A : Optional[Any]=1e-5 , A : Tuple=32 , A : Dict=None , A : int=None , **A : List[str] , ) -> Union[str, Any]: super().__init__(**A ) lowercase_ : Dict = image_size lowercase_ : int = patch_size lowercase_ : List[str] = num_channels lowercase_ : Union[str, Any] = embed_dim lowercase_ : Union[str, Any] = use_conv_embed lowercase_ : List[Any] = hidden_sizes lowercase_ : Union[str, Any] = depths lowercase_ : Tuple = focal_levels lowercase_ : List[Any] = focal_windows lowercase_ : Optional[Any] = hidden_act lowercase_ : List[Any] = mlp_ratio lowercase_ : Any = hidden_dropout_prob lowercase_ : str = drop_path_rate lowercase_ : int = use_layerscale lowercase_ : List[Any] = layerscale_value lowercase_ : Optional[Any] = use_post_layernorm lowercase_ : List[Any] = use_post_layernorm_in_modulation lowercase_ : int = normalize_modulator lowercase_ : Tuple = initializer_range lowercase_ : Tuple = layer_norm_eps lowercase_ : Tuple = encoder_stride lowercase_ : Union[str, Any] = ['''stem'''] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] lowercase_ , lowercase_ : Optional[int] = get_aligned_output_features_output_indices( out_features=A , out_indices=A , stage_names=self.stage_names )
33
"""simple docstring""" def lowercase ( __snake_case : Optional[int] ): lowercase_ : int = 0 lowercase_ : Optional[Any] = len(__snake_case ) for i in range(n - 1 ): for j in range(i + 1 , __snake_case ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowercase ( __snake_case : str ): if len(__snake_case ) <= 1: return arr, 0 lowercase_ : Optional[Any] = len(__snake_case ) // 2 lowercase_ : List[Any] = arr[0:mid] lowercase_ : Union[str, Any] = arr[mid:] lowercase_ , lowercase_ : Tuple = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = _count_cross_inversions(__snake_case , __snake_case ) lowercase_ : List[Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowercase ( __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Optional[Any] = [] lowercase_ : Any = 0 while i < len(__snake_case ) and j < len(__snake_case ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__snake_case ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__snake_case ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowercase ( ): lowercase_ : Union[str, Any] = [1_0, 2, 1, 5, 5, 2, 1_1] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowercase_ : int = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 8 print('''number of inversions = ''' , __snake_case ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowercase_ : Dict = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) # an empty list should also have zero inversions lowercase_ : List[Any] = [] lowercase_ : Any = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : List[str] = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) if __name__ == "__main__": main()
33
1
"""simple docstring""" import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase : def __init__( self : List[str] , A : Optional[Any] , A : Optional[int]=13 , A : List[str]=7 , A : str=True , A : List[Any]=True , A : List[str]=True , A : Dict=True , A : List[str]=99 , A : Any=24 , A : str=2 , A : Optional[Any]=6 , A : str=37 , A : List[Any]="gelu" , A : List[str]=0.1 , A : Union[str, Any]=0.1 , A : str=5_12 , A : List[Any]=16 , A : str=2 , A : Optional[int]=0.02 , A : List[Any]=3 , A : str=None , A : List[str]=10_00 , ) -> List[str]: lowercase_ : Union[str, Any] = parent lowercase_ : Optional[Any] = batch_size lowercase_ : int = seq_length lowercase_ : Union[str, Any] = is_training lowercase_ : Any = use_input_mask lowercase_ : Optional[int] = use_token_type_ids lowercase_ : List[str] = use_labels lowercase_ : Union[str, Any] = vocab_size lowercase_ : Any = hidden_size lowercase_ : Tuple = num_hidden_layers lowercase_ : str = num_attention_heads lowercase_ : Dict = intermediate_size lowercase_ : List[str] = hidden_act lowercase_ : Optional[Any] = hidden_dropout_prob lowercase_ : str = attention_probs_dropout_prob lowercase_ : str = max_position_embeddings lowercase_ : int = type_vocab_size lowercase_ : Optional[int] = type_sequence_label_size lowercase_ : Optional[Any] = initializer_range lowercase_ : List[Any] = num_labels lowercase_ : Dict = scope lowercase_ : int = range_bbox def A ( self : Optional[Any] ) -> Dict: lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : List[str] = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowercase_ : Optional[int] = bbox[i, j, 3] lowercase_ : List[str] = bbox[i, j, 1] lowercase_ : Optional[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: lowercase_ : Optional[Any] = bbox[i, j, 2] lowercase_ : Dict = bbox[i, j, 0] lowercase_ : List[Any] = t lowercase_ : Optional[int] = None if self.use_input_mask: lowercase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowercase_ : Dict = None if self.use_token_type_ids: lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase_ : Any = None lowercase_ : Optional[int] = None if self.use_labels: lowercase_ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : Tuple = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def A ( self : int ) -> Optional[int]: return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def A ( self : str , A : str , A : int , A : Dict , A : Optional[Any] , A : Any , A : Optional[int] , A : int , ) -> Tuple: lowercase_ : str = LiltModel(config=A ) model.to(A ) model.eval() lowercase_ : Any = model(A , bbox=A , attention_mask=A , token_type_ids=A ) lowercase_ : Dict = model(A , bbox=A , token_type_ids=A ) lowercase_ : Any = model(A , bbox=A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Optional[int] , A : List[Any] , A : Any , A : Dict , A : List[Any] , A : int , A : Tuple , A : Optional[Any] , ) -> int: lowercase_ : int = self.num_labels lowercase_ : Union[str, Any] = LiltForTokenClassification(config=A ) model.to(A ) model.eval() lowercase_ : Any = model( A , bbox=A , attention_mask=A , token_type_ids=A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , A : Any , A : Any , A : Dict , A : Optional[Any] , A : List[Any] , A : Optional[int] , A : List[str] , ) -> Any: lowercase_ : List[Any] = LiltForQuestionAnswering(config=A ) model.to(A ) model.eval() lowercase_ : List[Any] = model( A , bbox=A , attention_mask=A , token_type_ids=A , start_positions=A , end_positions=A , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : List[Any] ) -> Tuple: lowercase_ : str = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Tuple = config_and_inputs lowercase_ : List[Any] = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_torch class _UpperCAmelCase ( _A , _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ : Dict = ( { "feature-extraction": LiltModel, "question-answering": LiltForQuestionAnswering, "text-classification": LiltForSequenceClassification, "token-classification": LiltForTokenClassification, "zero-shot": LiltForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Any = False SCREAMING_SNAKE_CASE_ : Tuple = False def A ( self : int , A : Optional[Any] , A : Union[str, Any] , A : List[str] , A : Union[str, Any] , A : int ) -> str: return True def A ( self : Tuple ) -> List[Any]: lowercase_ : Dict = LiltModelTester(self ) lowercase_ : Tuple = ConfigTester(self , config_class=A , hidden_size=37 ) def A ( self : Any ) -> int: self.config_tester.run_common_tests() def A ( self : Dict ) -> Tuple: lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : str ) -> Dict: lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase_ : Dict = type self.model_tester.create_and_check_model(*A ) def A ( self : Any ) -> List[str]: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) def A ( self : List[Any] ) -> Any: lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) @slow def A ( self : int ) -> str: for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : List[Any] = LiltModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_torch @slow class _UpperCAmelCase ( unittest.TestCase ): def A ( self : str ) -> List[str]: lowercase_ : Any = LiltModel.from_pretrained('''SCUT-DLVCLab/lilt-roberta-en-base''' ).to(A ) lowercase_ : int = torch.tensor([[1, 2]] , device=A ) lowercase_ : Any = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=A ) # forward pass with torch.no_grad(): lowercase_ : Any = model(input_ids=A , bbox=A ) lowercase_ : Optional[int] = torch.Size([1, 2, 7_68] ) lowercase_ : str = torch.tensor( [[-0.0653, 0.0950, -0.0061], [-0.0545, 0.0926, -0.0324]] , device=A , ) self.assertTrue(outputs.last_hidden_state.shape , A ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , A , atol=1e-3 ) )
33
"""simple docstring""" __A : Any = { '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
33
1
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def lowercase ( __snake_case : List[Any] , __snake_case : Tuple=1 ): if n_shave_prefix_segments >= 0: return ".".join(path.split('''.''' )[n_shave_prefix_segments:] ) else: return ".".join(path.split('''.''' )[:n_shave_prefix_segments] ) def lowercase ( __snake_case : str , __snake_case : str=0 ): lowercase_ : Dict = [] for old_item in old_list: lowercase_ : Optional[Any] = old_item.replace('''in_layers.0''' , '''norm1''' ) lowercase_ : List[Any] = new_item.replace('''in_layers.2''' , '''conv1''' ) lowercase_ : Any = new_item.replace('''out_layers.0''' , '''norm2''' ) lowercase_ : int = new_item.replace('''out_layers.3''' , '''conv2''' ) lowercase_ : Any = new_item.replace('''emb_layers.1''' , '''time_emb_proj''' ) lowercase_ : Tuple = new_item.replace('''skip_connection''' , '''conv_shortcut''' ) lowercase_ : Union[str, Any] = shave_segments(__snake_case , n_shave_prefix_segments=__snake_case ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def lowercase ( __snake_case : Union[str, Any] , __snake_case : Optional[int]=0 ): lowercase_ : List[Any] = [] for old_item in old_list: lowercase_ : Optional[int] = old_item lowercase_ : Tuple = new_item.replace('''norm.weight''' , '''group_norm.weight''' ) lowercase_ : Dict = new_item.replace('''norm.bias''' , '''group_norm.bias''' ) lowercase_ : Optional[Any] = new_item.replace('''proj_out.weight''' , '''proj_attn.weight''' ) lowercase_ : Dict = new_item.replace('''proj_out.bias''' , '''proj_attn.bias''' ) lowercase_ : Tuple = shave_segments(__snake_case , n_shave_prefix_segments=__snake_case ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def lowercase ( __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Optional[int]=None , __snake_case : Union[str, Any]=None , __snake_case : Dict=None ): assert isinstance(__snake_case , __snake_case ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): lowercase_ : List[str] = old_checkpoint[path] lowercase_ : Union[str, Any] = old_tensor.shape[0] // 3 lowercase_ : List[str] = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) lowercase_ : Optional[Any] = old_tensor.shape[0] // config['''num_head_channels'''] // 3 lowercase_ : Optional[Any] = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) lowercase_ , lowercase_ , lowercase_ : Dict = old_tensor.split(channels // num_heads , dim=1 ) lowercase_ : int = query.reshape(__snake_case ) lowercase_ : Optional[int] = key.reshape(__snake_case ) lowercase_ : Optional[int] = value.reshape(__snake_case ) for path in paths: lowercase_ : Optional[int] = path['''new'''] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here lowercase_ : Optional[Any] = new_path.replace('''middle_block.0''' , '''mid_block.resnets.0''' ) lowercase_ : Optional[Any] = new_path.replace('''middle_block.1''' , '''mid_block.attentions.0''' ) lowercase_ : str = new_path.replace('''middle_block.2''' , '''mid_block.resnets.1''' ) if additional_replacements is not None: for replacement in additional_replacements: lowercase_ : List[str] = new_path.replace(replacement['''old'''] , replacement['''new'''] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: lowercase_ : List[Any] = old_checkpoint[path['''old''']][:, :, 0] else: lowercase_ : str = old_checkpoint[path['''old''']] def lowercase ( __snake_case : int , __snake_case : List[Any] ): lowercase_ : Any = {} lowercase_ : str = checkpoint['''time_embed.0.weight'''] lowercase_ : Tuple = checkpoint['''time_embed.0.bias'''] lowercase_ : Tuple = checkpoint['''time_embed.2.weight'''] lowercase_ : Dict = checkpoint['''time_embed.2.bias'''] lowercase_ : Tuple = checkpoint['''input_blocks.0.0.weight'''] lowercase_ : Union[str, Any] = checkpoint['''input_blocks.0.0.bias'''] lowercase_ : Dict = checkpoint['''out.0.weight'''] lowercase_ : Optional[int] = checkpoint['''out.0.bias'''] lowercase_ : Union[str, Any] = checkpoint['''out.2.weight'''] lowercase_ : Tuple = checkpoint['''out.2.bias'''] # Retrieves the keys for the input blocks only lowercase_ : Union[str, Any] = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''input_blocks''' in layer} ) lowercase_ : Tuple = { layer_id: [key for key in checkpoint if F'''input_blocks.{layer_id}''' in key] for layer_id in range(__snake_case ) } # Retrieves the keys for the middle blocks only lowercase_ : str = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''middle_block''' in layer} ) lowercase_ : Optional[int] = { layer_id: [key for key in checkpoint if F'''middle_block.{layer_id}''' in key] for layer_id in range(__snake_case ) } # Retrieves the keys for the output blocks only lowercase_ : Union[str, Any] = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''output_blocks''' in layer} ) lowercase_ : Optional[int] = { layer_id: [key for key in checkpoint if F'''output_blocks.{layer_id}''' in key] for layer_id in range(__snake_case ) } for i in range(1 , __snake_case ): lowercase_ : List[str] = (i - 1) // (config['''num_res_blocks'''] + 1) lowercase_ : str = (i - 1) % (config['''num_res_blocks'''] + 1) lowercase_ : List[Any] = [key for key in input_blocks[i] if F'''input_blocks.{i}.0''' in key] lowercase_ : List[str] = [key for key in input_blocks[i] if F'''input_blocks.{i}.1''' in key] if F'''input_blocks.{i}.0.op.weight''' in checkpoint: lowercase_ : Any = checkpoint[ F'''input_blocks.{i}.0.op.weight''' ] lowercase_ : Union[str, Any] = checkpoint[ F'''input_blocks.{i}.0.op.bias''' ] continue lowercase_ : int = renew_resnet_paths(__snake_case ) lowercase_ : List[str] = {'''old''': F'''input_blocks.{i}.0''', '''new''': F'''down_blocks.{block_id}.resnets.{layer_in_block_id}'''} lowercase_ : List[Any] = {'''old''': '''resnets.2.op''', '''new''': '''downsamplers.0.op'''} assign_to_checkpoint( __snake_case , __snake_case , __snake_case , additional_replacements=[meta_path, resnet_op] , config=__snake_case ) if len(__snake_case ): lowercase_ : List[str] = renew_attention_paths(__snake_case ) lowercase_ : Optional[Any] = { '''old''': F'''input_blocks.{i}.1''', '''new''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}''', } lowercase_ : Optional[int] = { F'''input_blocks.{i}.1.qkv.bias''': { '''key''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias''', '''query''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias''', '''value''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias''', }, F'''input_blocks.{i}.1.qkv.weight''': { '''key''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight''', '''query''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight''', '''value''': F'''down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight''', }, } assign_to_checkpoint( __snake_case , __snake_case , __snake_case , additional_replacements=[meta_path] , attention_paths_to_split=__snake_case , config=__snake_case , ) lowercase_ : List[Any] = middle_blocks[0] lowercase_ : List[str] = middle_blocks[1] lowercase_ : Union[str, Any] = middle_blocks[2] lowercase_ : Dict = renew_resnet_paths(__snake_case ) assign_to_checkpoint(__snake_case , __snake_case , __snake_case , config=__snake_case ) lowercase_ : Tuple = renew_resnet_paths(__snake_case ) assign_to_checkpoint(__snake_case , __snake_case , __snake_case , config=__snake_case ) lowercase_ : int = renew_attention_paths(__snake_case ) lowercase_ : Union[str, Any] = { '''middle_block.1.qkv.bias''': { '''key''': '''mid_block.attentions.0.key.bias''', '''query''': '''mid_block.attentions.0.query.bias''', '''value''': '''mid_block.attentions.0.value.bias''', }, '''middle_block.1.qkv.weight''': { '''key''': '''mid_block.attentions.0.key.weight''', '''query''': '''mid_block.attentions.0.query.weight''', '''value''': '''mid_block.attentions.0.value.weight''', }, } assign_to_checkpoint( __snake_case , __snake_case , __snake_case , attention_paths_to_split=__snake_case , config=__snake_case ) for i in range(__snake_case ): lowercase_ : List[str] = i // (config['''num_res_blocks'''] + 1) lowercase_ : Union[str, Any] = i % (config['''num_res_blocks'''] + 1) lowercase_ : int = [shave_segments(__snake_case , 2 ) for name in output_blocks[i]] lowercase_ : Dict = {} for layer in output_block_layers: lowercase_ , lowercase_ : Tuple = layer.split('''.''' )[0], shave_segments(__snake_case , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(__snake_case ) else: lowercase_ : List[Any] = [layer_name] if len(__snake_case ) > 1: lowercase_ : Any = [key for key in output_blocks[i] if F'''output_blocks.{i}.0''' in key] lowercase_ : Optional[Any] = [key for key in output_blocks[i] if F'''output_blocks.{i}.1''' in key] lowercase_ : int = renew_resnet_paths(__snake_case ) lowercase_ : Dict = renew_resnet_paths(__snake_case ) lowercase_ : Optional[Any] = {'''old''': F'''output_blocks.{i}.0''', '''new''': F'''up_blocks.{block_id}.resnets.{layer_in_block_id}'''} assign_to_checkpoint(__snake_case , __snake_case , __snake_case , additional_replacements=[meta_path] , config=__snake_case ) if ["conv.weight", "conv.bias"] in output_block_list.values(): lowercase_ : int = list(output_block_list.values() ).index(['''conv.weight''', '''conv.bias'''] ) lowercase_ : Any = checkpoint[ F'''output_blocks.{i}.{index}.conv.weight''' ] lowercase_ : Any = checkpoint[ F'''output_blocks.{i}.{index}.conv.bias''' ] # Clear attentions as they have been attributed above. if len(__snake_case ) == 2: lowercase_ : Dict = [] if len(__snake_case ): lowercase_ : int = renew_attention_paths(__snake_case ) lowercase_ : Union[str, Any] = { '''old''': F'''output_blocks.{i}.1''', '''new''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}''', } lowercase_ : str = { F'''output_blocks.{i}.1.qkv.bias''': { '''key''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias''', '''query''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias''', '''value''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias''', }, F'''output_blocks.{i}.1.qkv.weight''': { '''key''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight''', '''query''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight''', '''value''': F'''up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight''', }, } assign_to_checkpoint( __snake_case , __snake_case , __snake_case , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any('''qkv''' in key for key in attentions ) else None , config=__snake_case , ) else: lowercase_ : Dict = renew_resnet_paths(__snake_case , n_shave_prefix_segments=1 ) for path in resnet_0_paths: lowercase_ : int = '''.'''.join(['''output_blocks''', str(__snake_case ), path['''old''']] ) lowercase_ : Optional[int] = '''.'''.join(['''up_blocks''', str(__snake_case ), '''resnets''', str(__snake_case ), path['''new''']] ) lowercase_ : Tuple = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": __A : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') __A : Tuple = parser.parse_args() __A : Dict = torch.load(args.checkpoint_path) with open(args.config_file) as f: __A : Any = json.loads(f.read()) __A : str = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] __A : Optional[Any] = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: __A : Any = DDPMScheduler.from_config('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) __A : Optional[int] = VQModel.from_pretrained('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) __A : Tuple = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : List[Any] = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __A : Optional[Any] = logging.get_logger(__name__) __A : Optional[Any] = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __A : Tuple = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def lowercase ( __snake_case : Any , __snake_case : Optional[Any] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : int ): for attribute in key.split('''.''' ): lowercase_ : List[Any] = getattr(__snake_case , __snake_case ) if weight_type is not None: lowercase_ : Optional[int] = getattr(__snake_case , __snake_case ).shape else: lowercase_ : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": lowercase_ : List[Any] = value elif weight_type == "weight_g": lowercase_ : List[str] = value elif weight_type == "weight_v": lowercase_ : List[str] = value elif weight_type == "bias": lowercase_ : List[str] = value else: lowercase_ : List[Any] = value logger.info(F'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def lowercase ( __snake_case : str , __snake_case : Union[str, Any] ): lowercase_ : Optional[Any] = [] lowercase_ : int = fairseq_model.state_dict() lowercase_ : List[Any] = hf_model.feature_extractor lowercase_ : str = hf_model.adapter for name, value in fairseq_dict.items(): lowercase_ : int = False if "conv_layers" in name: load_conv_layer( __snake_case , __snake_case , __snake_case , __snake_case , hf_model.config.feat_extract_norm == '''group''' , ) lowercase_ : List[Any] = True elif any(x in name for x in ['''adaptor''', '''w2v_encoder.proj.''', '''w2v_proj_ln.'''] ): load_adapter(__snake_case , __snake_case , __snake_case , __snake_case ) lowercase_ : List[Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: lowercase_ : Optional[int] = True if "*" in mapped_key: lowercase_ : Any = name.split(__snake_case )[0].split('''.''' )[-2] lowercase_ : Any = mapped_key.replace('''*''' , __snake_case ) if "weight_g" in name: lowercase_ : Optional[int] = '''weight_g''' elif "weight_v" in name: lowercase_ : str = '''weight_v''' elif "bias" in name: lowercase_ : Optional[int] = '''bias''' elif "weight" in name: lowercase_ : Optional[Any] = '''weight''' else: lowercase_ : List[Any] = None set_recursively(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase ( __snake_case : Any , __snake_case : str , __snake_case : List[Any] , __snake_case : List[Any] , __snake_case : str ): lowercase_ : Union[str, Any] = full_name.split('''conv_layers.''' )[-1] lowercase_ : List[str] = name.split('''.''' ) lowercase_ : int = int(items[0] ) lowercase_ : Optional[int] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) lowercase_ : Optional[int] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) lowercase_ : Tuple = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) lowercase_ : List[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) lowercase_ : Optional[int] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__snake_case ) def lowercase ( __snake_case : Union[str, Any] , __snake_case : List[str] , __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Union[str, Any] = full_name.split('''adaptor.''' )[-1] lowercase_ : Optional[Any] = name.split('''.''' ) if items[1].isdigit(): lowercase_ : Optional[Any] = int(items[1] ) else: lowercase_ : Tuple = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.''' lowercase_ : int = value logger.info(F'''Adapter proj layer norm bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.''' lowercase_ : Tuple = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.''' lowercase_ : List[Any] = value logger.info(F'''Adapter proj layer bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.''' lowercase_ : Any = value logger.info(F'''Adapter proj layer weight was initialized from {full_name}.''' ) elif isinstance(__snake_case , __snake_case ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.''' lowercase_ : List[str] = value logger.info(F'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.''' lowercase_ : Any = value logger.info(F'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) else: unused_weights.append(__snake_case ) def lowercase ( __snake_case : int ): lowercase_ , lowercase_ : Any = emb.weight.shape lowercase_ : List[Any] = nn.Linear(__snake_case , __snake_case , bias=__snake_case ) lowercase_ : Optional[Any] = emb.weight.data return lin_layer @torch.no_grad() def lowercase ( __snake_case : Dict , __snake_case : List[Any] , __snake_case : List[Any] , __snake_case : Dict , __snake_case : Dict , __snake_case : Optional[int] , __snake_case : int , __snake_case : str , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : Any , ): lowercase_ : Any = WavaVecaConfig.from_pretrained( __snake_case , add_adapter=__snake_case , adapter_stride=__snake_case , adapter_kernel_size=__snake_case , use_auth_token=__snake_case , output_hidden_size=__snake_case , ) lowercase_ : Any = MBartConfig.from_pretrained(__snake_case ) # load model lowercase_ , lowercase_ , lowercase_ : List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ '''config_yaml''': config_yaml_path, '''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path, '''load_pretrained_decoder_from''': None, } , ) lowercase_ : str = model[0].eval() # load feature extractor lowercase_ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained(__snake_case , use_auth_token=__snake_case ) # set weights for wav2vec2 encoder lowercase_ : List[Any] = WavaVecaModel(__snake_case ) recursively_load_weights_wavaveca(model.encoder , __snake_case ) # load decoder weights lowercase_ : Optional[int] = MBartForCausalLM(__snake_case ) lowercase_ , lowercase_ : Dict = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=__snake_case ) logger.warning(F'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(F'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) lowercase_ : int = SpeechEncoderDecoderModel(encoder=__snake_case , decoder=__snake_case ) lowercase_ : Optional[int] = False lowercase_ : Optional[Any] = MBartaaTokenizer(__snake_case ) tokenizer.save_pretrained(__snake_case ) lowercase_ : List[str] = hf_wavavec.config.to_dict() lowercase_ : List[Any] = tokenizer.pad_token_id lowercase_ : Any = tokenizer.bos_token_id lowercase_ : str = tokenizer.eos_token_id lowercase_ : Dict = '''mbart50''' lowercase_ : Tuple = '''wav2vec2''' lowercase_ : Any = tokenizer.eos_token_id lowercase_ : int = 2_5_0_0_0_4 lowercase_ : Any = tokenizer.eos_token_id lowercase_ : Any = SpeechEncoderDecoderConfig.from_dict(__snake_case ) hf_wavavec.save_pretrained(__snake_case ) feature_extractor.save_pretrained(__snake_case ) if __name__ == "__main__": __A : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_yaml_path''', default=None, type=str, help='''Path to yaml file of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-xls-r-1b''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/mbart-large-50-one-to-many-mmt''', type=str, help='''Path to hf decoder checkpoint config''', ) parser.add_argument('''--add_adapter''', default=True, type=bool, help='''whethere to add model adapter layers''') parser.add_argument('''--adapter_stride''', default=2, type=int, help='''stride of adapter layers''') parser.add_argument('''--adapter_kernel_size''', default=3, type=int, help='''kernel size of adapter layers''') parser.add_argument('''--encoder_output_dim''', default=1_024, type=int, help='''encoder output dim''') parser.add_argument('''--start_token_id''', default=250_004, type=int, help='''`decoder_start_token_id` of model config''') __A : Union[str, Any] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
33
"""simple docstring""" import argparse import os import re import packaging.version __A : List[str] = '''examples/''' __A : int = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __A : Dict = { '''init''': '''src/transformers/__init__.py''', '''setup''': '''setup.py''', } __A : Optional[int] = '''README.md''' def lowercase ( __snake_case : int , __snake_case : Any , __snake_case : int ): with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : int = f.read() lowercase_ , lowercase_ : List[str] = REPLACE_PATTERNS[pattern] lowercase_ : Union[str, Any] = replace.replace('''VERSION''' , __snake_case ) lowercase_ : Optional[Any] = re_pattern.sub(__snake_case , __snake_case ) with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(__snake_case ) def lowercase ( __snake_case : int ): for folder, directories, fnames in os.walk(__snake_case ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__snake_case , __snake_case ) , __snake_case , pattern='''examples''' ) def lowercase ( __snake_case : Optional[Any] , __snake_case : Optional[Any]=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__snake_case , __snake_case , __snake_case ) if not patch: update_version_in_examples(__snake_case ) def lowercase ( ): lowercase_ : Union[str, Any] = '''🤗 Transformers currently provides the following architectures''' lowercase_ : Union[str, Any] = '''1. Want to contribute a new model?''' with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : List[str] = f.readlines() # Find the start of the list. lowercase_ : Optional[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase_ : str = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowercase_ : str = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__snake_case ) def lowercase ( ): with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowercase_ : List[Any] = f.read() lowercase_ : List[str] = REPLACE_PATTERNS['''init'''][0].search(__snake_case ).groups()[0] return packaging.version.parse(__snake_case ) def lowercase ( __snake_case : Optional[Any]=False ): lowercase_ : str = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowercase_ : Optional[Any] = default_version.base_version elif patch: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. lowercase_ : int = input(F'''Which version are you releasing? [{default_version}]''' ) if len(__snake_case ) == 0: lowercase_ : Dict = default_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case , patch=__snake_case ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def lowercase ( ): lowercase_ : List[Any] = get_version() lowercase_ : List[str] = F'''{current_version.major}.{current_version.minor + 1}.0.dev0''' lowercase_ : Any = current_version.base_version # Check with the user we got that right. lowercase_ : Tuple = input(F'''Which version are we developing now? [{dev_version}]''' ) if len(__snake_case ) == 0: lowercase_ : str = dev_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": __A : int = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __A : Any = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
33
1
"""simple docstring""" def lowercase ( __snake_case : int , __snake_case : int ): while b: lowercase_ , lowercase_ : str = b, a % b return a def lowercase ( __snake_case : int , __snake_case : int ): return a if b == 0 else euclidean_gcd_recursive(__snake_case , a % b ) def lowercase ( ): print(F'''euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}''' ) print(F'''euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}''' ) print(F'''euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}''' ) print(F'''euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}''' ) print(F'''euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}''' ) print(F'''euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}''' ) print(F'''euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}''' ) print(F'''euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}''' ) if __name__ == "__main__": main()
33
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def lowercase ( __snake_case : str , __snake_case : str , __snake_case : Optional[str] = None ): if version.parse(hfh.__version__ ).release < version.parse('''0.11.0''' ).release: # old versions of hfh don't url-encode the file path lowercase_ : Union[str, Any] = quote(__snake_case ) return hfh.hf_hub_url(__snake_case , __snake_case , repo_type='''dataset''' , revision=__snake_case )
33
1
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def lowercase ( __snake_case : List[Any] ): lowercase_ : int = os.path.join(args.tf_model_dir , '''parameters.json''' ) lowercase_ : Any = json.loads(open(__snake_case ).read() ) if not params: raise ValueError( F'''It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.''' ) if not args.output.endswith('''.pt''' ): lowercase_ : Dict = args.output + '''.pt''' lowercase_ : Any = OrderedDict() with tf.device('''/CPU:0''' ): lowercase_ : int = tf.train.load_checkpoint(args.tf_model_dir ) lowercase_ : Optional[int] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): lowercase_ : int = reader.get_tensor(__snake_case ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): lowercase_ : Optional[Any] = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): lowercase_ : Tuple = 8 lowercase_ : str = '''model.sqout.%d.weight''' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time lowercase_ : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : List[Any] = torch.tensor(__snake_case ) elif key_name.startswith('''model/moe''' ): lowercase_ : Optional[int] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): lowercase_ : List[str] = '''model.blocks.%d.feed_forward.mlp.router.classifier.weight''' % player lowercase_ : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : Tuple = torch.tensor(__snake_case ) elif key_name.endswith('''/softmlp/kernel''' ): lowercase_ : List[str] = '''model.blocks.%d.feed_forward.soft_bypass_mlp.weight''' % player lowercase_ : List[str] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : List[str] = torch.tensor(__snake_case ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): lowercase_ : Union[str, Any] = key_name[-9:-7] for i in range(1_6 ): lowercase_ : Tuple = '''model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight''' % (player, i, nlayer) lowercase_ : str = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided lowercase_ : Tuple = torch.tensor(__snake_case ) elif key_name.startswith('''model/mlp''' ): lowercase_ : Optional[int] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): lowercase_ : List[Any] = '''model.blocks.%d.feed_forward.mlp.wi.weight''' % player lowercase_ : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : Dict = torch.tensor(__snake_case ) elif key_name.endswith('''/p1/bias''' ): lowercase_ : Any = '''model.blocks.%d.feed_forward.mlp.wi.bias''' % player lowercase_ : Any = vnp.copy() # same because it is one dimensional lowercase_ : int = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/kernel''' ): lowercase_ : Optional[Any] = '''model.blocks.%d.feed_forward.mlp.wo.weight''' % player lowercase_ : int = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : int = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/bias''' ): lowercase_ : Optional[int] = '''model.blocks.%d.feed_forward.mlp.wo.bias''' % player lowercase_ : List[str] = vnp.copy() # same because it is one dimensional lowercase_ : int = torch.tensor(__snake_case ) elif key_name.startswith('''model/ln''' ): lowercase_ : Optional[Any] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): lowercase_ : str = '''model.blocks.%d.feed_forward.norm.bias''' % player lowercase_ : Any = vnp.copy() # same because it is one dimensional lowercase_ : int = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): lowercase_ : Union[str, Any] = '''model.blocks.%d.feed_forward.norm.weight''' % player lowercase_ : Union[str, Any] = vnp.copy() # same because it is one dimensional lowercase_ : int = torch.tensor(__snake_case ) elif key_name.startswith('''model/att''' ): lowercase_ : Optional[int] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): lowercase_ : Dict = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum lowercase_ : Tuple = state[:, 0, :, :] lowercase_ : Dict = state[:, 1, :, :] lowercase_ : Union[str, Any] = state[:, 2, :, :] lowercase_ : int = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase_ : Optional[Any] = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase_ : Union[str, Any] = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase_ : List[str] = '''model.blocks.%d.self_attn.self_attn.q_proj.weight''' % player lowercase_ : str = torch.tensor(__snake_case ) lowercase_ : str = '''model.blocks.%d.self_attn.self_attn.k_proj.weight''' % player lowercase_ : Any = torch.tensor(__snake_case ) lowercase_ : List[Any] = '''model.blocks.%d.self_attn.self_attn.v_proj.weight''' % player lowercase_ : Any = torch.tensor(__snake_case ) elif key_name.endswith('''/o/kernel''' ): lowercase_ : Dict = '''model.blocks.%d.self_attn.self_attn.out_proj.weight''' % player lowercase_ : Optional[int] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix lowercase_ : Any = torch.tensor(__snake_case ) elif key_name.startswith('''model/an''' ): lowercase_ : str = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): lowercase_ : Dict = '''model.blocks.%d.self_attn.norm.bias''' % player lowercase_ : Union[str, Any] = vnp.copy() # same because it is one dimensional lowercase_ : str = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): lowercase_ : str = '''model.blocks.%d.self_attn.norm.weight''' % player lowercase_ : Optional[Any] = vnp.copy() # same because it is one dimensional lowercase_ : Any = torch.tensor(__snake_case ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): lowercase_ : int = {'''wte''': '''embed_tokens''', '''wpe''': '''position_embeddings''', '''ete''': '''extra_position_embeddings'''}[ key_name[-3:] ] lowercase_ : str = '''model.%s.weight''' % nlayer lowercase_ : int = vnp.copy() # same in embedded lowercase_ : int = torch.tensor(__snake_case ) if key_name.startswith('''model/wte''' ): lowercase_ : Dict = '''lm_head.weight''' lowercase_ : Tuple = vnp.copy() # same in embedded lowercase_ : Dict = torch.tensor(__snake_case ) elif key_name.startswith('''model/wob''' ): lowercase_ : int = '''final_logits_bias''' lowercase_ : Any = vnp.copy() # same in embedded lowercase_ : Optional[Any] = state.reshape((1, -1) ) lowercase_ : Any = torch.tensor(__snake_case ) elif key_name == "model/dense/kernel": lowercase_ : List[str] = '''model.last_project.weight''' lowercase_ : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowercase_ : Optional[Any] = torch.tensor(__snake_case ) elif key_name == "model/dense_1/bias": lowercase_ : Dict = '''model.last_project.bias''' lowercase_ : Tuple = vnp.copy() # same because it is one dimensional lowercase_ : Any = torch.tensor(__snake_case ) torch.save(__snake_case , args.output ) if __name__ == "__main__": __A : Tuple = argparse.ArgumentParser( description='''model converter.''', formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument('''--tf_model_dir''', metavar='''PATH''', type=str, required=True, help='''import model''') parser.add_argument('''--output''', metavar='''PATH''', type=str, required=True, help='''output model''') __A : Any = parser.parse_args() convert_tf_gptsan_to_pt(args)
33
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self : int , A : Tuple , A : int=3 , A : List[str]=32 , A : Dict=3 , A : Any=10 , A : Dict=[10, 20, 30, 40] , A : Optional[Any]=[1, 1, 2, 1] , A : Union[str, Any]=True , A : Optional[Any]=True , A : Any="relu" , A : Optional[Any]=3 , A : Tuple=None , ) -> Dict: lowercase_ : str = parent lowercase_ : List[Any] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : int = num_channels lowercase_ : int = embeddings_size lowercase_ : str = hidden_sizes lowercase_ : List[str] = depths lowercase_ : Dict = is_training lowercase_ : int = use_labels lowercase_ : Any = hidden_act lowercase_ : List[Any] = num_labels lowercase_ : Tuple = scope lowercase_ : Optional[Any] = len(A ) def A ( self : str ) -> Tuple: lowercase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_labels ) lowercase_ : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Dict ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def A ( self : str , A : Tuple , A : str , A : str ) -> str: lowercase_ : str = TFResNetModel(config=A ) lowercase_ : Union[str, Any] = model(A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Any , A : int , A : List[Any] , A : Optional[Any] ) -> Optional[Any]: lowercase_ : Tuple = self.num_labels lowercase_ : Union[str, Any] = TFResNetForImageClassification(A ) lowercase_ : Tuple = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : Dict = config_and_inputs lowercase_ : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Any = False def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : int = TFResNetModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Dict ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Dict ) -> List[Any]: return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def A ( self : Any ) -> Any: pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def A ( self : List[str] ) -> Optional[Any]: pass def A ( self : str ) -> Tuple: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(A ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : str = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : List[str] ) -> Tuple: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : List[Any] ) -> List[str]: def check_hidden_states_output(A : Union[str, Any] , A : int , A : List[Any] ): lowercase_ : int = model_class(A ) lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Any = self.model_tester.num_stages self.assertEqual(len(A ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase_ : List[str] = layer_type lowercase_ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : Optional[Any] = True check_hidden_states_output(A , A , A ) def A ( self : Optional[int] ) -> Tuple: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def A ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFResNetModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase ( ): lowercase_ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Any ) -> Optional[int]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A ( self : Any ) -> Optional[int]: lowercase_ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase_ : List[Any] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : List[str] = image_processor(images=A , return_tensors='''tf''' ) # forward pass lowercase_ : Tuple = model(**A ) # verify the logits lowercase_ : Optional[int] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Optional[Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A , atol=1e-4 ) )
33
1
"""simple docstring""" from math import pow, sqrt def lowercase ( *__snake_case : float ): lowercase_ : Any = len(__snake_case ) > 0 and all(value > 0.0 for value in values ) return result def lowercase ( __snake_case : float , __snake_case : float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__snake_case , __snake_case ) else ValueError('''Input Error: Molar mass values must greater than 0.''' ) ) def lowercase ( __snake_case : float , __snake_case : float , __snake_case : float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__snake_case , __snake_case , __snake_case ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def lowercase ( __snake_case : float , __snake_case : float , __snake_case : float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(__snake_case , __snake_case , __snake_case ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def lowercase ( __snake_case : float , __snake_case : float , __snake_case : float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(__snake_case , __snake_case , __snake_case ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def lowercase ( __snake_case : float , __snake_case : float , __snake_case : float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(__snake_case , __snake_case , __snake_case ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) )
33
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin __A : Dict = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class _UpperCAmelCase ( unittest.TestCase , _A ): def A ( self : List[Any] ) -> Dict: lowercase_ : Optional[int] = load_tool('''text-question-answering''' ) self.tool.setup() lowercase_ : Union[str, Any] = load_tool('''text-question-answering''' , remote=A ) def A ( self : Any ) -> List[str]: lowercase_ : Union[str, Any] = self.tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : str ) -> List[str]: lowercase_ : int = self.remote_tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[Any] ) -> int: lowercase_ : Optional[Any] = self.tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[str] ) -> Optional[int]: lowercase_ : int = self.remote_tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' )
33
1
"""simple docstring""" from functools import lru_cache def lowercase ( __snake_case : int ): lowercase_ : Dict = 2 lowercase_ : Dict = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(__snake_case ) if n > 1: factors.add(__snake_case ) return factors @lru_cache def lowercase ( __snake_case : int ): return len(unique_prime_factors(__snake_case ) ) def lowercase ( __snake_case : list ): return len(set(__snake_case ) ) in (0, 1) def lowercase ( __snake_case : int ): lowercase_ : Dict = 2 while True: # Increment each value of a generated range lowercase_ : str = [base + i for i in range(__snake_case )] # Run elements through out unique_prime_factors function # Append our target number to the end. lowercase_ : Union[str, Any] = [upf_len(__snake_case ) for x in group] checker.append(__snake_case ) # If all numbers in the list are equal, return the group variable. if equality(__snake_case ): return group # Increment our base variable by 1 base += 1 def lowercase ( __snake_case : int = 4 ): lowercase_ : Optional[Any] = run(__snake_case ) return results[0] if len(__snake_case ) else None if __name__ == "__main__": print(solution())
33
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _UpperCAmelCase ( TensorFormatter[Mapping, "torch.Tensor", Mapping] ): def __init__( self : Any , A : int=None , **A : str ) -> Union[str, Any]: super().__init__(features=A ) lowercase_ : Union[str, Any] = torch_tensor_kwargs import torch # noqa import torch at initialization def A ( self : Dict , A : int ) -> List[Any]: import torch if isinstance(A , A ) and column: if all( isinstance(A , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(A ) return column def A ( self : int , A : Any ) -> Optional[Any]: import torch if isinstance(A , (str, bytes, type(A )) ): return value elif isinstance(A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() lowercase_ : Any = {} if isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): lowercase_ : Any = {'''dtype''': torch.intaa} elif isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): lowercase_ : Dict = {'''dtype''': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A , PIL.Image.Image ): lowercase_ : Dict = np.asarray(A ) return torch.tensor(A , **{**default_dtype, **self.torch_tensor_kwargs} ) def A ( self : Union[str, Any] , A : Optional[int] ) -> str: import torch # support for torch, tf, jax etc. if hasattr(A , '''__array__''' ) and not isinstance(A , torch.Tensor ): lowercase_ : Optional[int] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) elif isinstance(A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) return self._tensorize(A ) def A ( self : Dict , A : dict ) -> Tuple: return map_nested(self._recursive_tensorize , A , map_list=A ) def A ( self : str , A : pa.Table ) -> Mapping: lowercase_ : Optional[Any] = self.numpy_arrow_extractor().extract_row(A ) lowercase_ : str = self.python_features_decoder.decode_row(A ) return self.recursive_tensorize(A ) def A ( self : List[Any] , A : pa.Table ) -> "torch.Tensor": lowercase_ : List[str] = self.numpy_arrow_extractor().extract_column(A ) lowercase_ : str = self.python_features_decoder.decode_column(A , pa_table.column_names[0] ) lowercase_ : Optional[int] = self.recursive_tensorize(A ) lowercase_ : Any = self._consolidate(A ) return column def A ( self : List[str] , A : pa.Table ) -> Mapping: lowercase_ : Optional[int] = self.numpy_arrow_extractor().extract_batch(A ) lowercase_ : int = self.python_features_decoder.decode_batch(A ) lowercase_ : Dict = self.recursive_tensorize(A ) for column_name in batch: lowercase_ : Optional[Any] = self._consolidate(batch[column_name] ) return batch
33
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=_A ): SCREAMING_SNAKE_CASE_ : str = ["transformers", "torch", "note_seq"] def __init__( self : Tuple , *A : List[str] , **A : List[Any] ) -> str: requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def A ( cls : Union[str, Any] , *A : Any , **A : Union[str, Any] ) -> str: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def A ( cls : Tuple , *A : Union[str, Any] , **A : int ) -> Optional[int]: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
33
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
33
1
"""simple docstring""" import numpy class _UpperCAmelCase : def __init__( self : List[Any] , A : numpy.ndarray , A : numpy.ndarray ) -> None: lowercase_ : Union[str, Any] = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowercase_ : Optional[int] = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowercase_ : Optional[Any] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowercase_ : Optional[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowercase_ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowercase_ : Dict = numpy.zeros(output_array.shape ) def A ( self : Union[str, Any] ) -> numpy.ndarray: lowercase_ : List[Any] = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowercase_ : str = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowercase_ : Dict = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def A ( self : Optional[int] ) -> None: lowercase_ : Any = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowercase_ : int = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowercase_ : Dict = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def A ( self : str , A : numpy.ndarray , A : int , A : bool ) -> None: for iteration in range(1 , iterations + 1 ): lowercase_ : int = self.feedforward() self.back_propagation() if give_loss: lowercase_ : Optional[Any] = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F'''Iteration {iteration} Loss: {loss}''' ) def A ( self : Optional[Any] , A : numpy.ndarray ) -> int: lowercase_ : Optional[int] = input_arr lowercase_ : int = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowercase_ : Dict = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowercase_ : int = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowercase ( __snake_case : numpy.ndarray ): return 1 / (1 + numpy.exp(-value )) def lowercase ( __snake_case : numpy.ndarray ): return (value) * (1 - (value)) def lowercase ( ): lowercase_ : List[str] = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowercase_ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowercase_ : str = TwoHiddenLayerNeuralNetwork( input_array=__snake_case , output_array=__snake_case ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=__snake_case , iterations=1_0 , give_loss=__snake_case ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
33
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __A : Union[str, Any] = logging.get_logger(__name__) # General docstring __A : Tuple = '''MobileNetV1Config''' # Base docstring __A : Union[str, Any] = '''google/mobilenet_v1_1.0_224''' __A : Union[str, Any] = [1, 1_024, 7, 7] # Image classification docstring __A : Optional[Any] = '''google/mobilenet_v1_1.0_224''' __A : List[Any] = '''tabby, tabby cat''' __A : Union[str, Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=None ): lowercase_ : str = {} if isinstance(__snake_case , __snake_case ): lowercase_ : Union[str, Any] = model.mobilenet_va else: lowercase_ : Optional[Any] = model lowercase_ : Union[str, Any] = '''MobilenetV1/Conv2d_0/''' lowercase_ : Union[str, Any] = backbone.conv_stem.convolution.weight lowercase_ : Optional[Any] = backbone.conv_stem.normalization.bias lowercase_ : Union[str, Any] = backbone.conv_stem.normalization.weight lowercase_ : Any = backbone.conv_stem.normalization.running_mean lowercase_ : int = backbone.conv_stem.normalization.running_var for i in range(1_3 ): lowercase_ : Optional[int] = i + 1 lowercase_ : Union[str, Any] = i * 2 lowercase_ : Optional[Any] = backbone.layer[pt_index] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' lowercase_ : str = pointer.convolution.weight lowercase_ : int = pointer.normalization.bias lowercase_ : Any = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Union[str, Any] = pointer.normalization.running_var lowercase_ : Any = backbone.layer[pt_index + 1] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' lowercase_ : int = pointer.convolution.weight lowercase_ : str = pointer.normalization.bias lowercase_ : Tuple = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Any = pointer.normalization.running_var if isinstance(__snake_case , __snake_case ): lowercase_ : Optional[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase_ : Any = model.classifier.weight lowercase_ : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase_ : Tuple = tf.train.list_variables(__snake_case ) lowercase_ : int = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) lowercase_ : Optional[Any] = tf.train.load_variable(__snake_case , __snake_case ) lowercase_ : Optional[int] = array # Build TF to PyTorch weights loading map lowercase_ : Any = _build_tf_to_pytorch_map(__snake_case , __snake_case , __snake_case ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue lowercase_ : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase_ : Any = np.transpose(__snake_case , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase_ : Optional[int] = array.squeeze().transpose() else: lowercase_ : Optional[int] = np.transpose(__snake_case , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) lowercase_ : str = torch.from_numpy(__snake_case ) tf_weights.pop(__snake_case , __snake_case ) tf_weights.pop(name + '''/RMSProp''' , __snake_case ) tf_weights.pop(name + '''/RMSProp_1''' , __snake_case ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __snake_case ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def lowercase ( __snake_case : torch.Tensor , __snake_case : nn.Convad ): lowercase_ , lowercase_ : Optional[int] = features.shape[-2:] lowercase_ , lowercase_ : str = conv_layer.stride lowercase_ , lowercase_ : Tuple = conv_layer.kernel_size if in_height % stride_height == 0: lowercase_ : Dict = max(kernel_height - stride_height , 0 ) else: lowercase_ : List[Any] = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase_ : str = max(kernel_width - stride_width , 0 ) else: lowercase_ : int = max(kernel_width - (in_width % stride_width) , 0 ) lowercase_ : int = pad_along_width // 2 lowercase_ : Union[str, Any] = pad_along_width - pad_left lowercase_ : Tuple = pad_along_height // 2 lowercase_ : List[str] = pad_along_height - pad_top lowercase_ : str = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__snake_case , __snake_case , '''constant''' , 0.0 ) class _UpperCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : MobileNetVaConfig , A : int , A : int , A : int , A : Optional[int] = 1 , A : Optional[int] = 1 , A : bool = False , A : Optional[bool] = True , A : Optional[bool or str] = True , ) -> None: super().__init__() lowercase_ : int = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) lowercase_ : Tuple = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase_ : int = nn.Convad( in_channels=A , out_channels=A , kernel_size=A , stride=A , padding=A , groups=A , bias=A , padding_mode='''zeros''' , ) if use_normalization: lowercase_ : Optional[Any] = nn.BatchNormad( num_features=A , eps=config.layer_norm_eps , momentum=0.9997 , affine=A , track_running_stats=A , ) else: lowercase_ : Union[str, Any] = None if use_activation: if isinstance(A , A ): lowercase_ : str = ACTaFN[use_activation] elif isinstance(config.hidden_act , A ): lowercase_ : Any = ACTaFN[config.hidden_act] else: lowercase_ : Tuple = config.hidden_act else: lowercase_ : Tuple = None def A ( self : str , A : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: lowercase_ : List[Any] = apply_tf_padding(A , self.convolution ) lowercase_ : Optional[int] = self.convolution(A ) if self.normalization is not None: lowercase_ : Union[str, Any] = self.normalization(A ) if self.activation is not None: lowercase_ : Optional[int] = self.activation(A ) return features class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = MobileNetVaConfig SCREAMING_SNAKE_CASE_ : int = load_tf_weights_in_mobilenet_va SCREAMING_SNAKE_CASE_ : Optional[Any] = "mobilenet_v1" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any , A : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __A : Union[str, Any] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : str , A : MobileNetVaConfig , A : bool = True ) -> int: super().__init__(A ) lowercase_ : Union[str, Any] = config lowercase_ : List[str] = 32 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase_ : Union[str, Any] = MobileNetVaConvLayer( A , in_channels=config.num_channels , out_channels=A , kernel_size=3 , stride=2 , ) lowercase_ : Optional[Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase_ : List[Any] = nn.ModuleList() for i in range(13 ): lowercase_ : Dict = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=3 , stride=strides[i] , groups=A , ) ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=1 , ) ) lowercase_ : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def A ( self : Any , A : Optional[Any] ) -> Optional[int]: raise NotImplementedError @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : List[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase_ : List[str] = self.conv_stem(A ) lowercase_ : Dict = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase_ : Optional[int] = layer_module(A ) if output_hidden_states: lowercase_ : str = all_hidden_states + (hidden_states,) lowercase_ : Tuple = hidden_states if self.pooler is not None: lowercase_ : Dict = torch.flatten(self.pooler(A ) , start_dim=1 ) else: lowercase_ : Optional[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=A , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : List[str] , A : MobileNetVaConfig ) -> None: super().__init__(A ) lowercase_ : int = config.num_labels lowercase_ : List[str] = MobileNetVaModel(A ) lowercase_ : Union[str, Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase_ : Tuple = nn.Dropout(config.classifier_dropout_prob , inplace=A ) lowercase_ : int = nn.Linear(A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[Any] = self.mobilenet_va(A , output_hidden_states=A , return_dict=A ) lowercase_ : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Dict = self.classifier(self.dropout(A ) ) lowercase_ : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : Optional[Any] = '''single_label_classification''' else: lowercase_ : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : str = MSELoss() if self.num_labels == 1: lowercase_ : List[str] = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(A , A ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[Any] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : str = BCEWithLogitsLoss() lowercase_ : List[Any] = loss_fct(A , A ) if not return_dict: lowercase_ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A , logits=A , hidden_states=outputs.hidden_states , )
33
1
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __A : List[str] = logging.get_logger(__name__) __A : int = { '''ut/deta''': '''https://huggingface.co/ut/deta/resolve/main/config.json''', } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : int = "deta" SCREAMING_SNAKE_CASE_ : List[str] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : Union[str, Any] , A : Optional[int]=None , A : Union[str, Any]=9_00 , A : Tuple=20_48 , A : int=6 , A : str=20_48 , A : Any=8 , A : Optional[int]=6 , A : Dict=10_24 , A : str=8 , A : Dict=0.0 , A : Union[str, Any]=True , A : List[Any]="relu" , A : Tuple=2_56 , A : Optional[int]=0.1 , A : int=0.0 , A : str=0.0 , A : List[Any]=0.02 , A : Union[str, Any]=1.0 , A : str=True , A : str=False , A : Optional[int]="sine" , A : Optional[Any]=5 , A : str=4 , A : Union[str, Any]=4 , A : Tuple=True , A : Union[str, Any]=3_00 , A : Optional[Any]=True , A : int=True , A : Dict=1 , A : Tuple=5 , A : Optional[Any]=2 , A : Optional[Any]=1 , A : Any=1 , A : int=5 , A : Optional[Any]=2 , A : List[str]=0.1 , A : Dict=0.25 , **A : Tuple , ) -> Dict: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) lowercase_ : Optional[int] = CONFIG_MAPPING['''resnet'''](out_features=['''stage2''', '''stage3''', '''stage4'''] ) else: if isinstance(A , A ): lowercase_ : List[str] = backbone_config.pop('''model_type''' ) lowercase_ : List[str] = CONFIG_MAPPING[backbone_model_type] lowercase_ : Union[str, Any] = config_class.from_dict(A ) lowercase_ : List[str] = backbone_config lowercase_ : Optional[int] = num_queries lowercase_ : str = max_position_embeddings lowercase_ : Any = d_model lowercase_ : Optional[Any] = encoder_ffn_dim lowercase_ : List[str] = encoder_layers lowercase_ : Dict = encoder_attention_heads lowercase_ : int = decoder_ffn_dim lowercase_ : List[Any] = decoder_layers lowercase_ : int = decoder_attention_heads lowercase_ : Optional[Any] = dropout lowercase_ : Tuple = attention_dropout lowercase_ : str = activation_dropout lowercase_ : List[str] = activation_function lowercase_ : int = init_std lowercase_ : Dict = init_xavier_std lowercase_ : List[Any] = encoder_layerdrop lowercase_ : str = auxiliary_loss lowercase_ : Dict = position_embedding_type # deformable attributes lowercase_ : Union[str, Any] = num_feature_levels lowercase_ : Optional[int] = encoder_n_points lowercase_ : Dict = decoder_n_points lowercase_ : Tuple = two_stage lowercase_ : Union[str, Any] = two_stage_num_proposals lowercase_ : Tuple = with_box_refine lowercase_ : Optional[int] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''' ) # Hungarian matcher lowercase_ : Optional[Any] = class_cost lowercase_ : Dict = bbox_cost lowercase_ : Optional[int] = giou_cost # Loss coefficients lowercase_ : Optional[int] = mask_loss_coefficient lowercase_ : Optional[Any] = dice_loss_coefficient lowercase_ : Dict = bbox_loss_coefficient lowercase_ : int = giou_loss_coefficient lowercase_ : Union[str, Any] = eos_coefficient lowercase_ : Dict = focal_alpha super().__init__(is_encoder_decoder=A , **A ) @property def A ( self : Any ) -> int: return self.encoder_attention_heads @property def A ( self : Optional[int] ) -> int: return self.d_model def A ( self : List[Any] ) -> Dict: lowercase_ : str = copy.deepcopy(self.__dict__ ) lowercase_ : Union[str, Any] = self.backbone_config.to_dict() lowercase_ : List[Any] = self.__class__.model_type return output
33
"""simple docstring""" def lowercase ( __snake_case : list[int] ): lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): for j in range(i + 1 , __snake_case ): if numbers[j] < numbers[i]: lowercase_ , lowercase_ : Optional[int] = numbers[j], numbers[i] return numbers if __name__ == "__main__": __A : int = input('''Enter numbers separated by a comma:\n''').strip() __A : Any = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
33
1
"""simple docstring""" from math import ceil, sqrt def lowercase ( __snake_case : int = 1_0_0_0_0_0_0 ): lowercase_ : str = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: lowercase_ : Union[str, Any] = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: lowercase_ : Any = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F"""{solution() = }""")
33
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any] , A : Any , A : Tuple=7 , A : Tuple=3 , A : Optional[Any]=30 , A : List[Any]=4_00 , A : Tuple=True , A : Dict=None , A : List[str]=True , A : Optional[int]=[0.5, 0.5, 0.5] , A : Tuple=[0.5, 0.5, 0.5] , A : List[str]=True , A : List[Any]=1 / 2_55 , A : Union[str, Any]=True , ) -> Tuple: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase_ : Optional[int] = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} lowercase_ : Optional[int] = parent lowercase_ : str = batch_size lowercase_ : Tuple = num_channels lowercase_ : str = min_resolution lowercase_ : Any = max_resolution lowercase_ : str = do_resize lowercase_ : Any = size lowercase_ : Optional[int] = do_normalize lowercase_ : List[str] = image_mean lowercase_ : Optional[Any] = image_std lowercase_ : int = do_rescale lowercase_ : List[str] = rescale_factor lowercase_ : int = do_pad def A ( self : Any ) -> str: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Optional[Any] , A : int , A : int=False ) -> Tuple: if not batched: lowercase_ : Optional[int] = image_inputs[0] if isinstance(A , Image.Image ): lowercase_ , lowercase_ : int = image.size else: lowercase_ , lowercase_ : Tuple = image.shape[1], image.shape[2] if w < h: lowercase_ : int = int(self.size['''shortest_edge'''] * h / w ) lowercase_ : Optional[Any] = self.size['''shortest_edge'''] elif w > h: lowercase_ : Optional[Any] = self.size['''shortest_edge'''] lowercase_ : Optional[int] = int(self.size['''shortest_edge'''] * w / h ) else: lowercase_ : Any = self.size['''shortest_edge'''] lowercase_ : Any = self.size['''shortest_edge'''] else: lowercase_ : Tuple = [] for image in image_inputs: lowercase_ , lowercase_ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowercase_ : Union[str, Any] = max(A , key=lambda A : item[0] )[0] lowercase_ : Optional[Any] = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = YolosImageProcessor if is_vision_available() else None def A ( self : Optional[int] ) -> Optional[int]: lowercase_ : Optional[Any] = YolosImageProcessingTester(self ) @property def A ( self : str ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Optional[int] ) -> List[str]: lowercase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , '''image_mean''' ) ) self.assertTrue(hasattr(A , '''image_std''' ) ) self.assertTrue(hasattr(A , '''do_normalize''' ) ) self.assertTrue(hasattr(A , '''do_resize''' ) ) self.assertTrue(hasattr(A , '''size''' ) ) def A ( self : Dict ) -> Tuple: lowercase_ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowercase_ : Tuple = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A ) def A ( self : Optional[int] ) -> Tuple: pass def A ( self : Tuple ) -> int: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowercase_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ , lowercase_ : Dict = self.image_processor_tester.get_expected_values(A , batched=A ) lowercase_ : str = image_processing(A , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : str ) -> Any: # Initialize image_processing lowercase_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowercase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : int = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Optional[int] = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[Any] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[int]: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowercase_ : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Any = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[str] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[Any]: # Initialize image_processings lowercase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) lowercase_ : Tuple = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowercase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowercase_ : Union[str, Any] = image_processing_a.pad(A , return_tensors='''pt''' ) lowercase_ : List[Any] = image_processing_a(A , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1e-4 ) ) @slow def A ( self : str ) -> List[Any]: # prepare image and target lowercase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowercase_ : List[Any] = json.loads(f.read() ) lowercase_ : Tuple = {'''image_id''': 3_97_69, '''annotations''': target} # encode them lowercase_ : Union[str, Any] = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) lowercase_ : List[Any] = image_processing(images=A , annotations=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Union[str, Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Union[str, Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : Tuple = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : List[str] = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : Any = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify orig_size lowercase_ : List[str] = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : Optional[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) ) @slow def A ( self : List[Any] ) -> Dict: # prepare image, target and masks_path lowercase_ : Tuple = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowercase_ : str = json.loads(f.read() ) lowercase_ : int = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} lowercase_ : List[Any] = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowercase_ : int = YolosImageProcessor(format='''coco_panoptic''' ) lowercase_ : Any = image_processing(images=A , annotations=A , masks_path=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Optional[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Tuple = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : List[Any] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : str = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : List[str] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify masks lowercase_ : Dict = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A ) # verify orig_size lowercase_ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) )
33
1
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def lowercase ( __snake_case : Any , __snake_case : Dict ): if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer lowercase_ : List[str] = flax_key_tuple[:-1] + ('''weight''',) lowercase_ : List[str] = torch.permute(__snake_case , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(__snake_case ): # linear layer lowercase_ : List[str] = flax_key_tuple[:-1] + ('''weight''',) lowercase_ : Optional[Any] = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: lowercase_ : Union[str, Any] = flax_key_tuple[:-1] + ('''weight''',) return flax_key_tuple, flax_tensor def lowercase ( __snake_case : str , __snake_case : str , __snake_case : Dict ): if "metadata" in layer: lowercase_ : int = layer.split('''metadata''' ) lowercase_ : Union[str, Any] = ''''''.join(split_layer[0] )[:-1] lowercase_ : str = [tuple(('''metadata''' + split_layer[1]).split('''/''' ) )] elif "kvstore" in layer: lowercase_ : Union[str, Any] = layer.split('''kvstore''' ) lowercase_ : List[str] = ''''''.join(split_layer[0] )[:-1] lowercase_ : Any = [tuple(('''kvstore''' + split_layer[1]).split('''/''' ) )] else: lowercase_ : Union[str, Any] = layer.split('''/''' ) lowercase_ : Tuple = '''/'''.join(split_layer[:-1] ) lowercase_ : List[Any] = (split_layer[-1],) if "kvstore/path" in layer: lowercase_ : int = F'''{switch_checkpoint_path}/{checkpoint_info[layer]}''' elif "kvstore/driver" in layer: lowercase_ : Optional[Any] = '''file''' else: lowercase_ : Optional[Any] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def lowercase ( __snake_case : Any , __snake_case : Optional[Any] ): lowercase_ : str = rename_keys(__snake_case ) lowercase_ : int = {} for k, v in current_block.items(): lowercase_ : Tuple = v lowercase_ : List[Any] = new_current_block torch.save(__snake_case , __snake_case ) def lowercase ( __snake_case : List[Any] , __snake_case : Any , __snake_case : List[str] , __snake_case : Dict , __snake_case : str = WEIGHTS_NAME ): lowercase_ : Union[str, Any] = convert_file_size_to_int(__snake_case ) lowercase_ : Optional[Any] = [] lowercase_ : Tuple = {} lowercase_ : str = 0 lowercase_ : int = 0 os.makedirs(__snake_case , exist_ok=__snake_case ) with gfile.GFile(switch_checkpoint_path + '''/checkpoint''' , '''rb''' ) as fp: lowercase_ : int = serialization.msgpack_restore(fp.read() )['''optimizer''']['''target'''] lowercase_ : Tuple = flatten_dict(__snake_case , sep='''/''' ) lowercase_ : List[Any] = {} for layer in checkpoint_info.keys(): lowercase_ , lowercase_ , lowercase_ : Tuple = get_key_and_tensorstore_dict( __snake_case , __snake_case , __snake_case ) if curr_real_layer_name in all_layers: lowercase_ : Tuple = content else: lowercase_ : Optional[int] = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file lowercase_ : Optional[Any] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() lowercase_ : str = torch.tensor(__snake_case ) lowercase_ : Tuple = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts lowercase_ , lowercase_ : Dict = rename_base_flax_keys(tuple(key.split('''/''' ) ) , __snake_case ) lowercase_ : Union[str, Any] = '''/'''.join(__snake_case ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: lowercase_ : int = os.path.join( __snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) rename_and_save_block(__snake_case , __snake_case ) sharded_state_dicts.append(current_block.keys() ) del current_block lowercase_ : Dict = {} lowercase_ : Optional[int] = 0 lowercase_ : int = raw_weights.to(getattr(__snake_case , __snake_case ) ) current_block_size += weight_size total_size += weight_size # Add the last block lowercase_ : List[Any] = os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) rename_and_save_block(__snake_case , __snake_case ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(__snake_case ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index lowercase_ : Any = {} lowercase_ : Dict = {} for idx, shard in enumerate(__snake_case ): lowercase_ : Union[str, Any] = weights_name.replace( '''.bin''' , F'''-{idx+1:05d}-of-{len(__snake_case ):05d}.bin''' ) # len(sharded_state_dicts):05d} lowercase_ : Tuple = os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(__snake_case , os.path.join(__snake_case , __snake_case ) ) lowercase_ : List[str] = shard for key in shard: lowercase_ : Union[str, Any] = shard_file # Add the metadata lowercase_ : int = {'''total_size''': total_size} lowercase_ : Tuple = {'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(__snake_case , __snake_case ) , '''w''' , encoding='''utf-8''' ) as f: lowercase_ : Dict = json.dumps(__snake_case , indent=2 , sort_keys=__snake_case ) + '''\n''' f.write(__snake_case ) return metadata, index if __name__ == "__main__": __A : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--switch_t5x_checkpoint_path''', default='''/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600''', type=str, required=False, help='''Path to a directory containing a folder per layer. Follows the original Google format.''', ) parser.add_argument('''--max_shard_size''', default='''10GB''', required=False, help='''Max shard size''') parser.add_argument('''--dtype''', default='''bfloat16''', type=str, required=False, help='''dtype of the saved model''') parser.add_argument( '''--pytorch_dump_folder_path''', default='''/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted''', type=str, required=False, help='''Path to the output pytorch model.''', ) __A : str = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def lowercase ( ): from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer lowercase_ : List[str] = SwitchTransformersConfig.from_pretrained('''google/switch-base-8''' ) config.save_pretrained('''/home/arthur_huggingface_co/transformers/switch_converted''' ) lowercase_ : Optional[Any] = SwitchTransformersForConditionalGeneration.from_pretrained( '''/home/arthur_huggingface_co/transformers/switch_converted''' , device_map='''auto''' ) lowercase_ : Union[str, Any] = TaTokenizer.from_pretrained('''t5-small''' ) lowercase_ : List[Any] = '''A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''' lowercase_ : List[str] = tokenizer(__snake_case , return_tensors='''pt''' ).input_ids lowercase_ : List[Any] = model.generate(__snake_case , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0 ): lowercase_ : str = 0 lowercase_ : List[Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
33
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : int = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : Any = ( { "feature-extraction": TFMobileBertModel, "fill-mask": TFMobileBertForMaskedLM, "question-answering": TFMobileBertForQuestionAnswering, "text-classification": TFMobileBertForSequenceClassification, "token-classification": TFMobileBertForTokenClassification, "zero-shot": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : int = False def A ( self : Dict , A : Dict , A : List[Any] , A : Tuple=False ) -> int: lowercase_ : str = super()._prepare_for_class(A , A , return_labels=A ) if return_labels: if model_class in get_values(A ): lowercase_ : List[str] = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _A ): def __init__( self : Any , A : int , A : int=13 , A : Any=7 , A : Optional[Any]=True , A : List[str]=True , A : List[Any]=True , A : Dict=True , A : List[str]=99 , A : str=32 , A : str=32 , A : List[Any]=2 , A : Tuple=4 , A : Optional[Any]=37 , A : Tuple="gelu" , A : Optional[int]=0.1 , A : Tuple=0.1 , A : Optional[int]=5_12 , A : List[Any]=16 , A : Optional[Any]=2 , A : Any=0.02 , A : List[str]=3 , A : Dict=4 , A : Tuple=None , ) -> List[Any]: lowercase_ : Optional[int] = parent lowercase_ : Optional[int] = batch_size lowercase_ : Union[str, Any] = seq_length lowercase_ : List[str] = is_training lowercase_ : str = use_input_mask lowercase_ : Tuple = use_token_type_ids lowercase_ : Any = use_labels lowercase_ : Optional[int] = vocab_size lowercase_ : Tuple = hidden_size lowercase_ : List[str] = num_hidden_layers lowercase_ : Dict = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Tuple = hidden_act lowercase_ : Optional[Any] = hidden_dropout_prob lowercase_ : Optional[Any] = attention_probs_dropout_prob lowercase_ : str = max_position_embeddings lowercase_ : str = type_vocab_size lowercase_ : Tuple = type_sequence_label_size lowercase_ : str = initializer_range lowercase_ : str = num_labels lowercase_ : Optional[Any] = num_choices lowercase_ : Any = scope lowercase_ : int = embedding_size def A ( self : List[Any] ) -> Any: lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : Any = None if self.use_input_mask: lowercase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : int = None if self.use_token_type_ids: lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase_ : str = None lowercase_ : Dict = None lowercase_ : Any = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : int = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ : Optional[Any] = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] , A : List[str] , A : Optional[int] , A : str , A : Union[str, Any] , A : List[str] , A : Optional[Any] , A : Union[str, Any] ) -> Optional[Any]: lowercase_ : Optional[int] = TFMobileBertModel(config=A ) lowercase_ : List[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Tuple = model(A ) lowercase_ : Optional[int] = [input_ids, input_mask] lowercase_ : int = model(A ) lowercase_ : Optional[int] = model(A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : List[Any] , A : int , A : int , A : int , A : int , A : Optional[int] , A : List[str] , A : Dict ) -> List[Any]: lowercase_ : List[Any] = TFMobileBertForMaskedLM(config=A ) lowercase_ : Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , A : Optional[Any] , A : int , A : Dict , A : Union[str, Any] , A : List[Any] , A : List[str] , A : List[Any] ) -> Union[str, Any]: lowercase_ : Optional[int] = TFMobileBertForNextSentencePrediction(config=A ) lowercase_ : int = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def A ( self : Any , A : List[str] , A : Union[str, Any] , A : Any , A : Any , A : Tuple , A : Optional[Any] , A : str ) -> List[str]: lowercase_ : Optional[Any] = TFMobileBertForPreTraining(config=A ) lowercase_ : List[str] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : List[Any] = model(A ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def A ( self : Dict , A : Optional[int] , A : List[Any] , A : List[Any] , A : Union[str, Any] , A : Optional[Any] , A : Any , A : str ) -> str: lowercase_ : Tuple = self.num_labels lowercase_ : Optional[Any] = TFMobileBertForSequenceClassification(config=A ) lowercase_ : Optional[int] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : List[str] , A : Optional[int] , A : Union[str, Any] , A : int , A : Optional[int] , A : List[Any] , A : Any , A : Tuple ) -> str: lowercase_ : Tuple = self.num_choices lowercase_ : List[str] = TFMobileBertForMultipleChoice(config=A ) lowercase_ : Tuple = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Dict = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : List[str] = tf.tile(tf.expand_dims(A , 1 ) , (1, self.num_choices, 1) ) lowercase_ : Optional[int] = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } lowercase_ : Union[str, Any] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Tuple , A : Optional[int] , A : Optional[Any] , A : Any , A : Dict , A : Optional[int] , A : Dict , A : str ) -> str: lowercase_ : Optional[Any] = self.num_labels lowercase_ : Optional[int] = TFMobileBertForTokenClassification(config=A ) lowercase_ : List[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Tuple = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Dict , A : Dict , A : Optional[Any] , A : List[Any] , A : Union[str, Any] , A : List[str] , A : int , A : str ) -> Optional[Any]: lowercase_ : Optional[int] = TFMobileBertForQuestionAnswering(config=A ) lowercase_ : str = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : Union[str, Any] = model(A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : int ) -> Union[str, Any]: lowercase_ : Optional[int] = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Any = config_and_inputs lowercase_ : List[str] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict def A ( self : Tuple ) -> Tuple: lowercase_ : List[Any] = TFMobileBertModelTest.TFMobileBertModelTester(self ) lowercase_ : List[str] = ConfigTester(self , config_class=A , hidden_size=37 ) def A ( self : int ) -> Optional[Any]: self.config_tester.run_common_tests() def A ( self : List[str] ) -> Tuple: lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*A ) def A ( self : List[Any] ) -> Union[str, Any]: lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*A ) def A ( self : List[str] ) -> List[Any]: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*A ) def A ( self : str ) -> Any: lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*A ) def A ( self : Optional[int] ) -> int: lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*A ) def A ( self : str ) -> int: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*A ) def A ( self : Optional[Any] ) -> List[str]: lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*A ) def A ( self : str ) -> Any: lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*A ) @slow def A ( self : Optional[Any] ) -> Optional[Any]: # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: lowercase_ : Any = TFMobileBertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : Any = TFMobileBertForPreTraining.from_pretrained('''google/mobilebert-uncased''' ) lowercase_ : Any = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase_ : List[Any] = model(A )[0] lowercase_ : Optional[int] = [1, 6, 3_05_22] self.assertEqual(output.shape , A ) lowercase_ : Tuple = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A , atol=1e-4 )
33
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
1
"""simple docstring""" from __future__ import annotations import bisect def lowercase ( __snake_case : list[int] , __snake_case : int , __snake_case : int = 0 , __snake_case : int = -1 ): if hi < 0: lowercase_ : List[Any] = len(__snake_case ) while lo < hi: lowercase_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: lowercase_ : str = mid + 1 else: lowercase_ : List[str] = mid return lo def lowercase ( __snake_case : list[int] , __snake_case : int , __snake_case : int = 0 , __snake_case : int = -1 ): if hi < 0: lowercase_ : Any = len(__snake_case ) while lo < hi: lowercase_ : Dict = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: lowercase_ : Optional[int] = mid + 1 else: lowercase_ : Any = mid return lo def lowercase ( __snake_case : list[int] , __snake_case : int , __snake_case : int = 0 , __snake_case : int = -1 ): sorted_collection.insert(bisect_left(__snake_case , __snake_case , __snake_case , __snake_case ) , __snake_case ) def lowercase ( __snake_case : list[int] , __snake_case : int , __snake_case : int = 0 , __snake_case : int = -1 ): sorted_collection.insert(bisect_right(__snake_case , __snake_case , __snake_case , __snake_case ) , __snake_case ) def lowercase ( __snake_case : list[int] , __snake_case : int ): lowercase_ : Optional[int] = 0 lowercase_ : Tuple = len(__snake_case ) - 1 while left <= right: lowercase_ : str = left + (right - left) // 2 lowercase_ : Dict = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: lowercase_ : Dict = midpoint - 1 else: lowercase_ : Tuple = midpoint + 1 return None def lowercase ( __snake_case : list[int] , __snake_case : int ): lowercase_ : int = bisect.bisect_left(__snake_case , __snake_case ) if index != len(__snake_case ) and sorted_collection[index] == item: return index return None def lowercase ( __snake_case : list[int] , __snake_case : int , __snake_case : int , __snake_case : int ): if right < left: return None lowercase_ : List[str] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(__snake_case , __snake_case , __snake_case , midpoint - 1 ) else: return binary_search_by_recursion(__snake_case , __snake_case , midpoint + 1 , __snake_case ) if __name__ == "__main__": __A : Dict = input('''Enter numbers separated by comma:\n''').strip() __A : List[str] = sorted(int(item) for item in user_input.split(''',''')) __A : Optional[int] = int(input('''Enter a single number to be found in the list:\n''')) __A : int = binary_search(collection, target) if result is None: print(F"""{target} was not found in {collection}.""") else: print(F"""{target} was found at position {result} in {collection}.""")
33
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
1
"""simple docstring""" import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class _UpperCAmelCase : def __init__( self : Union[str, Any] , A : Tuple , A : Optional[Any]=13 , A : int=10 , A : Dict=3 , A : Union[str, Any]=2 , A : Dict=2 , A : Tuple=2 , A : str=True , A : str=True , A : List[str]=32 , A : Optional[int]=5 , A : Any=4 , A : Dict=37 , A : Optional[int]="gelu" , A : List[Any]=0.1 , A : List[str]=0.1 , A : Optional[Any]=10 , A : Optional[int]=0.02 , A : Any=0.9 , A : List[Any]=None , ) -> Any: lowercase_ : Optional[int] = parent lowercase_ : Dict = batch_size lowercase_ : Optional[int] = image_size lowercase_ : Optional[Any] = num_channels lowercase_ : Tuple = patch_size lowercase_ : int = tubelet_size lowercase_ : Union[str, Any] = num_frames lowercase_ : List[str] = is_training lowercase_ : List[Any] = use_labels lowercase_ : List[str] = hidden_size lowercase_ : Dict = num_hidden_layers lowercase_ : Any = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Any = hidden_act lowercase_ : Optional[Any] = hidden_dropout_prob lowercase_ : Tuple = attention_probs_dropout_prob lowercase_ : Optional[int] = type_sequence_label_size lowercase_ : int = initializer_range lowercase_ : List[str] = mask_ratio lowercase_ : Optional[Any] = scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame lowercase_ : Optional[int] = (image_size // patch_size) ** 2 lowercase_ : int = (num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos lowercase_ : Tuple = int(mask_ratio * self.seq_length ) def A ( self : str ) -> str: lowercase_ : List[Any] = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Optional[int] = None if self.use_labels: lowercase_ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def A ( self : Optional[Any] ) -> List[Any]: return VideoMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , tubelet_size=self.tubelet_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] , A : Optional[int] , A : str , A : Optional[int] ) -> Any: lowercase_ : Union[str, Any] = VideoMAEModel(config=A ) model.to(A ) model.eval() lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Dict , A : str , A : str , A : Any ) -> Any: lowercase_ : List[Any] = VideoMAEForPreTraining(A ) model.to(A ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowercase_ : Union[str, Any] = torch.ones((self.num_masks,) ) lowercase_ : Tuple = torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) lowercase_ : List[Any] = mask.expand(self.batch_size , -1 ).bool() lowercase_ : Tuple = model(A , A ) # model only returns predictions for masked patches lowercase_ : List[str] = mask.sum().item() lowercase_ : Any = 3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_masked_patches, decoder_num_labels) ) def A ( self : List[Any] ) -> Any: lowercase_ : Union[str, Any] = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : str = config_and_inputs lowercase_ : Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : List[str] = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ : str = ( {"feature-extraction": VideoMAEModel, "video-classification": VideoMAEForVideoClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Any = False SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any ) -> int: lowercase_ : int = VideoMAEModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A , hidden_size=37 ) def A ( self : List[Any] , A : Optional[int] , A : Optional[int] , A : str=False ) -> str: lowercase_ : List[str] = copy.deepcopy(A ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowercase_ : Union[str, Any] = torch.ones((self.model_tester.num_masks,) ) lowercase_ : List[str] = torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) lowercase_ : Optional[Any] = mask.expand(self.model_tester.batch_size , -1 ).bool() lowercase_ : Union[str, Any] = bool_masked_pos.to(A ) if return_labels: if model_class in [ *get_values(A ), ]: lowercase_ : Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A ) return inputs_dict def A ( self : Any ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason='''VideoMAE does not use inputs_embeds''' ) def A ( self : Dict ) -> List[Any]: pass def A ( self : Tuple ) -> str: lowercase_ , lowercase_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : Dict = model_class(A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase_ : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A , nn.Linear ) ) def A ( self : Union[str, Any] ) -> Optional[Any]: lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : Union[str, Any] = model_class(A ) lowercase_ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : Optional[Any] = [*signature.parameters.keys()] lowercase_ : Dict = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : Tuple ) -> List[str]: lowercase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : Dict ) -> int: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*A ) @slow def A ( self : Optional[Any] ) -> Optional[Any]: for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Union[str, Any] = VideoMAEModel.from_pretrained(A ) self.assertIsNotNone(A ) def A ( self : Union[str, Any] ) -> int: if not self.has_attentions: pass else: lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : List[str] = True for model_class in self.all_model_classes: lowercase_ : Dict = self.model_tester.seq_length - self.model_tester.num_masks lowercase_ : Union[str, Any] = ( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) lowercase_ : Tuple = True lowercase_ : Union[str, Any] = False lowercase_ : Dict = True lowercase_ : Any = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : List[str] = outputs.attentions self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowercase_ : Optional[int] = True lowercase_ : Union[str, Any] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): lowercase_ : Dict = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.attentions self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) lowercase_ : Any = len(A ) # Check attention is always last and order is fine lowercase_ : List[str] = True lowercase_ : Optional[Any] = True lowercase_ : List[Any] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): lowercase_ : List[Any] = model(**self._prepare_for_class(A , A ) ) self.assertEqual(out_len + 1 , len(A ) ) lowercase_ : Union[str, Any] = outputs.attentions self.assertEqual(len(A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def A ( self : int ) -> List[str]: def check_hidden_states_output(A : Dict , A : List[Any] , A : Optional[Any] ): lowercase_ : Tuple = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): lowercase_ : Union[str, Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Any = outputs.hidden_states lowercase_ : Any = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(A ) , A ) lowercase_ : Optional[int] = self.model_tester.seq_length - self.model_tester.num_masks lowercase_ : List[Any] = num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : List[str] = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : List[Any] = True check_hidden_states_output(A , A , A ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def A ( self : Tuple ) -> Tuple: pass def lowercase ( ): lowercase_ : Optional[int] = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) lowercase_ : str = np.load(__snake_case ) return list(__snake_case ) @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : List[Any] ) -> List[str]: # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def A ( self : List[Any] ) -> str: lowercase_ : List[str] = VideoMAEForVideoClassification.from_pretrained('''MCG-NJU/videomae-base-finetuned-kinetics''' ).to( A ) lowercase_ : str = self.default_image_processor lowercase_ : List[str] = prepare_video() lowercase_ : Union[str, Any] = image_processor(A , return_tensors='''pt''' ).to(A ) # forward pass with torch.no_grad(): lowercase_ : List[str] = model(**A ) # verify the logits lowercase_ : Tuple = torch.Size((1, 4_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Union[str, Any] = torch.tensor([0.3669, -0.0688, -0.2421] ).to(A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A , atol=1e-4 ) ) @slow def A ( self : List[Any] ) -> List[Any]: lowercase_ : int = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ).to(A ) lowercase_ : Any = self.default_image_processor lowercase_ : Union[str, Any] = prepare_video() lowercase_ : Tuple = image_processor(A , return_tensors='''pt''' ).to(A ) # add boolean mask, indicating which patches to mask lowercase_ : int = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' , filename='''bool_masked_pos.pt''' ) lowercase_ : Any = torch.load(A ) # forward pass with torch.no_grad(): lowercase_ : List[Any] = model(**A ) # verify the logits lowercase_ : Tuple = torch.Size([1, 14_08, 15_36] ) lowercase_ : Tuple = torch.tensor( [[0.7994, 0.9612, 0.8508], [0.7401, 0.8958, 0.8302], [0.5862, 0.7468, 0.7325]] , device=A ) self.assertEqual(outputs.logits.shape , A ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , A , atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) lowercase_ : Any = torch.tensor([0.5142] , device=A ) self.assertTrue(torch.allclose(outputs.loss , A , atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) lowercase_ : Tuple = VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' , norm_pix_loss=A ).to( A ) with torch.no_grad(): lowercase_ : int = model(**A ) lowercase_ : Dict = torch.tensor(torch.tensor([0.6469] ) , device=A ) self.assertTrue(torch.allclose(outputs.loss , A , atol=1e-4 ) )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0_0 ): lowercase_ , lowercase_ : str = 1, 1 lowercase_ : List[str] = 2 while True: lowercase_ : Tuple = 0 lowercase_ : List[Any] = fa + fa lowercase_ , lowercase_ : Optional[int] = fa, f index += 1 for _ in str(__snake_case ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A : List[Any] = { '''configuration_roc_bert''': ['''ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoCBertConfig'''], '''tokenization_roc_bert''': ['''RoCBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ '''ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoCBertForCausalLM''', '''RoCBertForMaskedLM''', '''RoCBertForMultipleChoice''', '''RoCBertForPreTraining''', '''RoCBertForQuestionAnswering''', '''RoCBertForSequenceClassification''', '''RoCBertForTokenClassification''', '''RoCBertLayer''', '''RoCBertModel''', '''RoCBertPreTrainedModel''', '''load_tf_weights_in_roc_bert''', ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : Union[str, Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = "vit_mae" def __init__( self : Dict , A : List[str]=7_68 , A : Any=12 , A : Union[str, Any]=12 , A : Tuple=30_72 , A : Any="gelu" , A : Tuple=0.0 , A : List[str]=0.0 , A : Tuple=0.02 , A : Tuple=1e-12 , A : int=2_24 , A : Dict=16 , A : int=3 , A : Tuple=True , A : Tuple=16 , A : Optional[Any]=5_12 , A : Union[str, Any]=8 , A : List[Any]=20_48 , A : Dict=0.75 , A : Any=False , **A : Optional[int] , ) -> Union[str, Any]: super().__init__(**A ) lowercase_ : List[Any] = hidden_size lowercase_ : str = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : List[Any] = hidden_dropout_prob lowercase_ : int = attention_probs_dropout_prob lowercase_ : int = initializer_range lowercase_ : Dict = layer_norm_eps lowercase_ : Optional[Any] = image_size lowercase_ : str = patch_size lowercase_ : Dict = num_channels lowercase_ : Any = qkv_bias lowercase_ : Union[str, Any] = decoder_num_attention_heads lowercase_ : Optional[Any] = decoder_hidden_size lowercase_ : List[str] = decoder_num_hidden_layers lowercase_ : List[Any] = decoder_intermediate_size lowercase_ : Optional[Any] = mask_ratio lowercase_ : Optional[Any] = norm_pix_loss
33
1
"""simple docstring""" from __future__ import annotations def lowercase ( __snake_case : tuple[int, int] , __snake_case : int ): lowercase_ , lowercase_ : int = position lowercase_ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowercase_ : Union[str, Any] = [] for position in positions: lowercase_ , lowercase_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(__snake_case ) return permissible_positions def lowercase ( __snake_case : list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def lowercase ( __snake_case : list[list[int]] , __snake_case : tuple[int, int] , __snake_case : int ): if is_complete(__snake_case ): return True for position in get_valid_pos(__snake_case , len(__snake_case ) ): lowercase_ , lowercase_ : str = position if board[y][x] == 0: lowercase_ : Optional[Any] = curr + 1 if open_knight_tour_helper(__snake_case , __snake_case , curr + 1 ): return True lowercase_ : Optional[int] = 0 return False def lowercase ( __snake_case : int ): lowercase_ : Optional[Any] = [[0 for i in range(__snake_case )] for j in range(__snake_case )] for i in range(__snake_case ): for j in range(__snake_case ): lowercase_ : Dict = 1 if open_knight_tour_helper(__snake_case , (i, j) , 1 ): return board lowercase_ : Optional[int] = 0 lowercase_ : int = F'''Open Kight Tour cannot be performed on a board of size {n}''' raise ValueError(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" def lowercase ( __snake_case : int ): if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: lowercase_ : Dict = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase ( __snake_case : int ): lowercase_ : str = 0 lowercase_ : List[str] = 2 while digits < n: index += 1 lowercase_ : Any = len(str(fibonacci(__snake_case ) ) ) return index def lowercase ( __snake_case : int = 1_0_0_0 ): return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0_0_0_0_0 ): lowercase_ : Union[str, Any] = set(range(3 , __snake_case , 2 ) ) primes.add(2 ) for p in range(3 , __snake_case , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , __snake_case , __snake_case ) ) ) lowercase_ : Dict = [float(__snake_case ) for n in range(limit + 1 )] for p in primes: for n in range(__snake_case , limit + 1 , __snake_case ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"""{solution() = }""")
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __A : List[str] = { '''configuration_mobilenet_v2''': [ '''MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileNetV2Config''', '''MobileNetV2OnnxConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''MobileNetV2FeatureExtractor'''] __A : Optional[int] = ['''MobileNetV2ImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileNetV2ForImageClassification''', '''MobileNetV2ForSemanticSegmentation''', '''MobileNetV2Model''', '''MobileNetV2PreTrainedModel''', '''load_tf_weights_in_mobilenet_v2''', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : Union[str, Any] ) -> Dict: lowercase_ : List[str] = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) lowercase_ : List[str] = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house lowercase_ : Dict = torch.Size((1, 12, 7_68) ) # batch_size, sequence_length, embedding_vector_dim lowercase_ : int = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): lowercase_ : str = model(A )['''last_hidden_state'''].detach() self.assertEqual(output.shape , A ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , A , atol=1e-3 ) ) @slow def A ( self : Optional[int] ) -> List[str]: lowercase_ : List[Any] = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) lowercase_ : Union[str, Any] = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house lowercase_ : Optional[Any] = torch.Size((1, 12, 10_24) ) # batch_size, sequence_length, embedding_vector_dim lowercase_ : List[str] = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): lowercase_ : Optional[int] = model(A )['''last_hidden_state'''].detach() self.assertEqual(output.shape , A ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , A , atol=1e-3 ) )
33
"""simple docstring""" from __future__ import annotations __A : List[Any] = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __A : str = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): lowercase_ : float = -1 for j in range(i + 1 , __snake_case ): if arr[i] < arr[j]: lowercase_ : List[str] = arr[j] break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] for i, outer in enumerate(__snake_case ): lowercase_ : float = -1 for inner in arr[i + 1 :]: if outer < inner: lowercase_ : List[Any] = inner break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = len(__snake_case ) lowercase_ : list[float] = [] lowercase_ : list[float] = [-1] * arr_size for index in reversed(range(__snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowercase_ : Optional[Any] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __A : int = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
33
1
"""simple docstring""" import numpy as np from PIL import Image def lowercase ( __snake_case : np.ndarray , __snake_case : int , __snake_case : int ): lowercase_ : Tuple = np.array(__snake_case ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) lowercase_ : Dict = 0 lowercase_ : Any = 0 lowercase_ : List[str] = 0 lowercase_ : Union[str, Any] = 0 # compute the shape of the output matrix lowercase_ : Optional[int] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape lowercase_ : Union[str, Any] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix lowercase_ : Tuple = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 lowercase_ : Any = 0 lowercase_ : Optional[Any] = 0 return updated_arr def lowercase ( __snake_case : np.ndarray , __snake_case : int , __snake_case : int ): lowercase_ : int = np.array(__snake_case ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) lowercase_ : int = 0 lowercase_ : Dict = 0 lowercase_ : Tuple = 0 lowercase_ : Tuple = 0 # compute the shape of the output matrix lowercase_ : List[str] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape lowercase_ : Optional[int] = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix lowercase_ : str = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 lowercase_ : int = 0 lowercase_ : str = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='''avgpooling''', verbose=True) # Loading the image __A : List[Any] = Image.open('''path_to_image''') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __A : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowercase ( ): lowercase_ : Union[str, Any] = { '''repo_name''': ['''test_repo1''', '''test_repo2''', '''test_repo3'''], '''path''': ['''test_1.py''', '''test_2.py''', '''unit_test.py'''], '''content''': ['''a ''' * 2_0, '''a ''' * 3_0, '''b ''' * 7], } lowercase_ : List[Any] = Dataset.from_dict(__snake_case ) return dataset class _UpperCAmelCase ( _A ): def A ( self : Any ) -> Optional[Any]: lowercase_ : Optional[Any] = get_dataset() lowercase_ : int = make_duplicate_clusters(A , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def A ( self : Optional[int] ) -> int: lowercase_ : List[Any] = get_dataset() lowercase_ , lowercase_ : Optional[Any] = deduplicate_dataset(A ) self.assertEqual(len(A ) , 2 ) print(A ) self.assertEqual(duplicate_clusters[0][0]['''copies'''] , 2 ) self.assertEqual(duplicate_clusters[0][0]['''is_extreme'''] , A )
33
"""simple docstring""" def lowercase ( __snake_case : int ): if not isinstance(__snake_case , __snake_case ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) __A : Optional[int] = { '''configuration_speech_to_text''': ['''SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Speech2TextConfig'''], '''processing_speech_to_text''': ['''Speech2TextProcessor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = ['''Speech2TextTokenizer'''] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = ['''Speech2TextFeatureExtractor'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = [ '''TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSpeech2TextForConditionalGeneration''', '''TFSpeech2TextModel''', '''TFSpeech2TextPreTrainedModel''', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = [ '''SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Speech2TextForConditionalGeneration''', '''Speech2TextModel''', '''Speech2TextPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys __A : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" def lowercase ( __snake_case : Optional[int] ): lowercase_ : int = 0 lowercase_ : Optional[Any] = len(__snake_case ) for i in range(n - 1 ): for j in range(i + 1 , __snake_case ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowercase ( __snake_case : str ): if len(__snake_case ) <= 1: return arr, 0 lowercase_ : Optional[Any] = len(__snake_case ) // 2 lowercase_ : List[Any] = arr[0:mid] lowercase_ : Union[str, Any] = arr[mid:] lowercase_ , lowercase_ : Tuple = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = _count_cross_inversions(__snake_case , __snake_case ) lowercase_ : List[Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowercase ( __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Optional[Any] = [] lowercase_ : Any = 0 while i < len(__snake_case ) and j < len(__snake_case ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__snake_case ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__snake_case ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowercase ( ): lowercase_ : Union[str, Any] = [1_0, 2, 1, 5, 5, 2, 1_1] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowercase_ : int = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 8 print('''number of inversions = ''' , __snake_case ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowercase_ : Dict = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) # an empty list should also have zero inversions lowercase_ : List[Any] = [] lowercase_ : Any = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : List[str] = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) if __name__ == "__main__": main()
33
1
"""simple docstring""" __A : List[Any] = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' __A : Dict = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __A : int = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
33
"""simple docstring""" __A : Any = { '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
33
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class _UpperCAmelCase : def __init__( self : Optional[int] , A : List[str] , A : Optional[Any]=13 , A : Dict=7 , A : Optional[int]=True , A : Tuple=True , A : Optional[Any]=True , A : Any=True , A : str=99 , A : Dict=32 , A : List[str]=2 , A : Tuple=4 , A : Any=37 , A : Tuple="gelu" , A : Any=0.1 , A : int=0.1 , A : Optional[int]=5_12 , A : Optional[int]=16 , A : List[Any]=2 , A : Union[str, Any]=0.02 , A : int=False , A : int=True , A : Dict="None" , A : Dict=3 , A : List[Any]=4 , A : int=None , ) -> Optional[int]: lowercase_ : Tuple = parent lowercase_ : Dict = batch_size lowercase_ : Union[str, Any] = seq_length lowercase_ : Dict = is_training lowercase_ : Any = use_input_mask lowercase_ : Dict = use_token_type_ids lowercase_ : Dict = use_labels lowercase_ : List[Any] = vocab_size lowercase_ : Any = hidden_size lowercase_ : Tuple = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Union[str, Any] = intermediate_size lowercase_ : str = hidden_act lowercase_ : List[Any] = hidden_dropout_prob lowercase_ : Union[str, Any] = attention_probs_dropout_prob lowercase_ : Union[str, Any] = max_position_embeddings lowercase_ : Optional[Any] = type_vocab_size lowercase_ : Any = type_sequence_label_size lowercase_ : int = initializer_range lowercase_ : Any = num_labels lowercase_ : int = num_choices lowercase_ : List[str] = relative_attention lowercase_ : str = position_biased_input lowercase_ : str = pos_att_type lowercase_ : int = scope def A ( self : Dict ) -> List[str]: lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : Any = None if self.use_input_mask: lowercase_ : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : str = None if self.use_token_type_ids: lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase_ : List[str] = None lowercase_ : List[Any] = None lowercase_ : Tuple = None if self.use_labels: lowercase_ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : Union[str, Any] = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=A , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Tuple , A : Tuple , A : Dict , A : Dict , A : List[Any] , A : str , A : List[Any] , A : List[Any] ) -> List[Any]: lowercase_ : List[str] = TFDebertaVaModel(config=A ) lowercase_ : Optional[int] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase_ : List[str] = [input_ids, input_mask] lowercase_ : List[Any] = model(A ) lowercase_ : str = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : int , A : Tuple , A : Any , A : Any , A : int , A : Dict , A : List[str] , A : int ) -> List[Any]: lowercase_ : str = TFDebertaVaForMaskedLM(config=A ) lowercase_ : int = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : Tuple = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Union[str, Any] , A : Any , A : Optional[Any] , A : Optional[int] , A : Tuple , A : str , A : Tuple , A : Optional[Any] ) -> Union[str, Any]: lowercase_ : Dict = self.num_labels lowercase_ : Tuple = TFDebertaVaForSequenceClassification(config=A ) lowercase_ : Union[str, Any] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Tuple , A : Any , A : Any , A : Tuple , A : Optional[int] , A : List[str] , A : Optional[Any] , A : List[Any] ) -> Optional[int]: lowercase_ : Any = self.num_labels lowercase_ : Optional[int] = TFDebertaVaForTokenClassification(config=A ) lowercase_ : List[Any] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : str = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Optional[Any] , A : Any , A : Any , A : List[Any] , A : Any , A : Dict , A : List[Any] , A : Union[str, Any] ) -> List[str]: lowercase_ : Any = TFDebertaVaForQuestionAnswering(config=A ) lowercase_ : List[str] = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : List[Any] ) -> int: lowercase_ : List[Any] = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : List[str] = config_and_inputs lowercase_ : Tuple = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : List[Any] = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : int = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[Any] = False SCREAMING_SNAKE_CASE_ : Tuple = False def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Any = TFDebertaVaModelTester(self ) lowercase_ : Dict = ConfigTester(self , config_class=A , hidden_size=37 ) def A ( self : List[Any] ) -> Union[str, Any]: self.config_tester.run_common_tests() def A ( self : int ) -> Dict: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : Optional[int] ) -> str: lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A ) def A ( self : Optional[Any] ) -> Any: lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) def A ( self : Optional[Any] ) -> List[Any]: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A ) def A ( self : Tuple ) -> Optional[Any]: lowercase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) @slow def A ( self : Tuple ) -> List[Any]: lowercase_ : Any = TFDebertaVaModel.from_pretrained('''kamalkraj/deberta-v2-xlarge''' ) self.assertIsNotNone(A ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): @unittest.skip(reason='''Model not available yet''' ) def A ( self : Dict ) -> Dict: pass @slow def A ( self : List[str] ) -> Union[str, Any]: lowercase_ : List[str] = TFDebertaVaModel.from_pretrained('''kamalkraj/deberta-v2-xlarge''' ) lowercase_ : Tuple = tf.constant([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) lowercase_ : Dict = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowercase_ : Union[str, Any] = model(A , attention_mask=A )[0] lowercase_ : int = tf.constant( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , A , atol=1e-4 )
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : List[Any] = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" import re from filelock import FileLock try: import nltk __A : List[Any] = True except (ImportError, ModuleNotFoundError): __A : List[str] = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def lowercase ( __snake_case : str ): re.sub('''<n>''' , '''''' , __snake_case ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__snake_case ) )
33
"""simple docstring""" import argparse import os import re import packaging.version __A : List[str] = '''examples/''' __A : int = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __A : Dict = { '''init''': '''src/transformers/__init__.py''', '''setup''': '''setup.py''', } __A : Optional[int] = '''README.md''' def lowercase ( __snake_case : int , __snake_case : Any , __snake_case : int ): with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : int = f.read() lowercase_ , lowercase_ : List[str] = REPLACE_PATTERNS[pattern] lowercase_ : Union[str, Any] = replace.replace('''VERSION''' , __snake_case ) lowercase_ : Optional[Any] = re_pattern.sub(__snake_case , __snake_case ) with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(__snake_case ) def lowercase ( __snake_case : int ): for folder, directories, fnames in os.walk(__snake_case ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__snake_case , __snake_case ) , __snake_case , pattern='''examples''' ) def lowercase ( __snake_case : Optional[Any] , __snake_case : Optional[Any]=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__snake_case , __snake_case , __snake_case ) if not patch: update_version_in_examples(__snake_case ) def lowercase ( ): lowercase_ : Union[str, Any] = '''🤗 Transformers currently provides the following architectures''' lowercase_ : Union[str, Any] = '''1. Want to contribute a new model?''' with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : List[str] = f.readlines() # Find the start of the list. lowercase_ : Optional[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase_ : str = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowercase_ : str = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__snake_case ) def lowercase ( ): with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowercase_ : List[Any] = f.read() lowercase_ : List[str] = REPLACE_PATTERNS['''init'''][0].search(__snake_case ).groups()[0] return packaging.version.parse(__snake_case ) def lowercase ( __snake_case : Optional[Any]=False ): lowercase_ : str = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowercase_ : Optional[Any] = default_version.base_version elif patch: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. lowercase_ : int = input(F'''Which version are you releasing? [{default_version}]''' ) if len(__snake_case ) == 0: lowercase_ : Dict = default_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case , patch=__snake_case ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def lowercase ( ): lowercase_ : List[Any] = get_version() lowercase_ : List[str] = F'''{current_version.major}.{current_version.minor + 1}.0.dev0''' lowercase_ : Any = current_version.base_version # Check with the user we got that right. lowercase_ : Tuple = input(F'''Which version are we developing now? [{dev_version}]''' ) if len(__snake_case ) == 0: lowercase_ : str = dev_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": __A : int = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __A : Any = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
33
1
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class _UpperCAmelCase : def __init__( self : Any , A : int ) -> None: lowercase_ : List[str] = value lowercase_ : Node | None = None lowercase_ : Node | None = None class _UpperCAmelCase : def __init__( self : Optional[int] , A : Node ) -> None: lowercase_ : Optional[Any] = tree def A ( self : Any , A : Node | None ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : int ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def lowercase ( __snake_case : str , __snake_case : str , __snake_case : Optional[str] = None ): if version.parse(hfh.__version__ ).release < version.parse('''0.11.0''' ).release: # old versions of hfh don't url-encode the file path lowercase_ : Union[str, Any] = quote(__snake_case ) return hfh.hf_hub_url(__snake_case , __snake_case , repo_type='''dataset''' , revision=__snake_case )
33
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __A : Dict = logging.get_logger(__name__) class _UpperCAmelCase ( _A ): def __init__( self : int , *A : Tuple , **A : int ) -> None: warnings.warn( '''The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use VideoMAEImageProcessor instead.''' , A , ) super().__init__(*A , **A )
33
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self : int , A : Tuple , A : int=3 , A : List[str]=32 , A : Dict=3 , A : Any=10 , A : Dict=[10, 20, 30, 40] , A : Optional[Any]=[1, 1, 2, 1] , A : Union[str, Any]=True , A : Optional[Any]=True , A : Any="relu" , A : Optional[Any]=3 , A : Tuple=None , ) -> Dict: lowercase_ : str = parent lowercase_ : List[Any] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : int = num_channels lowercase_ : int = embeddings_size lowercase_ : str = hidden_sizes lowercase_ : List[str] = depths lowercase_ : Dict = is_training lowercase_ : int = use_labels lowercase_ : Any = hidden_act lowercase_ : List[Any] = num_labels lowercase_ : Tuple = scope lowercase_ : Optional[Any] = len(A ) def A ( self : str ) -> Tuple: lowercase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_labels ) lowercase_ : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Dict ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def A ( self : str , A : Tuple , A : str , A : str ) -> str: lowercase_ : str = TFResNetModel(config=A ) lowercase_ : Union[str, Any] = model(A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Any , A : int , A : List[Any] , A : Optional[Any] ) -> Optional[Any]: lowercase_ : Tuple = self.num_labels lowercase_ : Union[str, Any] = TFResNetForImageClassification(A ) lowercase_ : Tuple = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : Dict = config_and_inputs lowercase_ : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Any = False def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : int = TFResNetModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Dict ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Dict ) -> List[Any]: return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def A ( self : Any ) -> Any: pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def A ( self : List[str] ) -> Optional[Any]: pass def A ( self : str ) -> Tuple: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(A ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : str = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : List[str] ) -> Tuple: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : List[Any] ) -> List[str]: def check_hidden_states_output(A : Union[str, Any] , A : int , A : List[Any] ): lowercase_ : int = model_class(A ) lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Any = self.model_tester.num_stages self.assertEqual(len(A ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase_ : List[str] = layer_type lowercase_ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : Optional[Any] = True check_hidden_states_output(A , A , A ) def A ( self : Optional[int] ) -> Tuple: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def A ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFResNetModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase ( ): lowercase_ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Any ) -> Optional[int]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A ( self : Any ) -> Optional[int]: lowercase_ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase_ : List[Any] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : List[str] = image_processor(images=A , return_tensors='''tf''' ) # forward pass lowercase_ : Tuple = model(**A ) # verify the logits lowercase_ : Optional[int] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Optional[Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A , atol=1e-4 ) )
33
1
"""simple docstring""" from __future__ import annotations def lowercase ( __snake_case : str , __snake_case : list[str] | None = None ): lowercase_ : Dict = word_bank or [] # create a table lowercase_ : int = len(__snake_case ) + 1 lowercase_ : list[list[list[str]]] = [] for _ in range(__snake_case ): table.append([] ) # seed value lowercase_ : str = [[]] # because empty string has empty combination # iterate through the indices for i in range(__snake_case ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(__snake_case )] == word: lowercase_ : list[list[str]] = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(__snake_case )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(__snake_case )]: combination.reverse() return table[len(__snake_case )] if __name__ == "__main__": print(all_construct('''jwajalapa''', ['''jwa''', '''j''', '''w''', '''a''', '''la''', '''lapa'''])) print(all_construct('''rajamati''', ['''s''', '''raj''', '''amat''', '''raja''', '''ma''', '''i''', '''t'''])) print( all_construct( '''hexagonosaurus''', ['''h''', '''ex''', '''hex''', '''ag''', '''ago''', '''ru''', '''auru''', '''rus''', '''go''', '''no''', '''o''', '''s'''], ) )
33
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin __A : Dict = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class _UpperCAmelCase ( unittest.TestCase , _A ): def A ( self : List[Any] ) -> Dict: lowercase_ : Optional[int] = load_tool('''text-question-answering''' ) self.tool.setup() lowercase_ : Union[str, Any] = load_tool('''text-question-answering''' , remote=A ) def A ( self : Any ) -> List[str]: lowercase_ : Union[str, Any] = self.tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : str ) -> List[str]: lowercase_ : int = self.remote_tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[Any] ) -> int: lowercase_ : Optional[Any] = self.tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[str] ) -> Optional[int]: lowercase_ : int = self.remote_tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' )
33
1
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def lowercase ( __snake_case : int ): random.seed(__snake_case ) np.random.seed(__snake_case ) torch.manual_seed(__snake_case ) torch.cuda.manual_seed_all(__snake_case ) # ^^ safe to call this function even if cuda is not available class _UpperCAmelCase : def __init__( self : Dict , A : Iterable[torch.nn.Parameter] , A : float = 0.9999 , A : float = 0.0 , A : int = 0 , A : bool = False , A : Union[float, int] = 1.0 , A : Union[float, int] = 2 / 3 , A : Optional[Any] = None , A : Dict[str, Any] = None , **A : str , ) -> Union[str, Any]: if isinstance(A , torch.nn.Module ): lowercase_ : List[Any] = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage`''' , '''1.0.0''' , A , standard_warn=A , ) lowercase_ : int = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility lowercase_ : Dict = True if kwargs.get('''max_value''' , A ) is not None: lowercase_ : Tuple = '''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('''max_value''' , '''1.0.0''' , A , standard_warn=A ) lowercase_ : List[str] = kwargs['''max_value'''] if kwargs.get('''min_value''' , A ) is not None: lowercase_ : Union[str, Any] = '''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('''min_value''' , '''1.0.0''' , A , standard_warn=A ) lowercase_ : int = kwargs['''min_value'''] lowercase_ : Union[str, Any] = list(A ) lowercase_ : int = [p.clone().detach() for p in parameters] if kwargs.get('''device''' , A ) is not None: lowercase_ : Union[str, Any] = '''The `device` argument is deprecated. Please use `to` instead.''' deprecate('''device''' , '''1.0.0''' , A , standard_warn=A ) self.to(device=kwargs['''device'''] ) lowercase_ : Optional[Any] = None lowercase_ : Optional[int] = decay lowercase_ : Optional[int] = min_decay lowercase_ : List[Any] = update_after_step lowercase_ : int = use_ema_warmup lowercase_ : Optional[int] = inv_gamma lowercase_ : Dict = power lowercase_ : str = 0 lowercase_ : Any = None # set in `step()` lowercase_ : List[str] = model_cls lowercase_ : int = model_config @classmethod def A ( cls : int , A : int , A : Optional[int] ) -> "EMAModel": lowercase_ , lowercase_ : List[str] = model_cls.load_config(A , return_unused_kwargs=A ) lowercase_ : Optional[int] = model_cls.from_pretrained(A ) lowercase_ : List[str] = cls(model.parameters() , model_cls=A , model_config=model.config ) ema_model.load_state_dict(A ) return ema_model def A ( self : Optional[Any] , A : List[str] ) -> Any: if self.model_cls is None: raise ValueError('''`save_pretrained` can only be used if `model_cls` was defined at __init__.''' ) if self.model_config is None: raise ValueError('''`save_pretrained` can only be used if `model_config` was defined at __init__.''' ) lowercase_ : Union[str, Any] = self.model_cls.from_config(self.model_config ) lowercase_ : Dict = self.state_dict() state_dict.pop('''shadow_params''' , A ) model.register_to_config(**A ) self.copy_to(model.parameters() ) model.save_pretrained(A ) def A ( self : str , A : int ) -> float: lowercase_ : Any = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: lowercase_ : Any = 1 - (1 + step / self.inv_gamma) ** -self.power else: lowercase_ : Optional[Any] = (1 + step) / (10 + step) lowercase_ : str = min(A , self.decay ) # make sure decay is not smaller than min_decay lowercase_ : Any = max(A , self.min_decay ) return cur_decay_value @torch.no_grad() def A ( self : Optional[int] , A : Iterable[torch.nn.Parameter] ) -> Dict: if isinstance(A , torch.nn.Module ): lowercase_ : str = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage.step`''' , '''1.0.0''' , A , standard_warn=A , ) lowercase_ : Optional[int] = parameters.parameters() lowercase_ : List[Any] = list(A ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. lowercase_ : Union[str, Any] = self.get_decay(self.optimization_step ) lowercase_ : Tuple = decay lowercase_ : Optional[int] = 1 - decay lowercase_ : Optional[Any] = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , A ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): lowercase_ : Tuple = deepspeed.zero.GatheredParameters(A , modifier_rank=A ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(A ) def A ( self : Tuple , A : Iterable[torch.nn.Parameter] ) -> None: lowercase_ : int = list(A ) for s_param, param in zip(self.shadow_params , A ): param.data.copy_(s_param.to(param.device ).data ) def A ( self : Any , A : Optional[int]=None , A : Optional[int]=None ) -> None: lowercase_ : Union[str, Any] = [ p.to(device=A , dtype=A ) if p.is_floating_point() else p.to(device=A ) for p in self.shadow_params ] def A ( self : Union[str, Any] ) -> dict: return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def A ( self : Tuple , A : Iterable[torch.nn.Parameter] ) -> None: lowercase_ : Tuple = [param.detach().cpu().clone() for param in parameters] def A ( self : Union[str, Any] , A : Iterable[torch.nn.Parameter] ) -> None: if self.temp_stored_params is None: raise RuntimeError('''This ExponentialMovingAverage has no `store()`ed weights ''' '''to `restore()`''' ) for c_param, param in zip(self.temp_stored_params , A ): param.data.copy_(c_param.data ) # Better memory-wise. lowercase_ : int = None def A ( self : Tuple , A : dict ) -> None: lowercase_ : Dict = copy.deepcopy(A ) lowercase_ : List[Any] = state_dict.get('''decay''' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('''Decay must be between 0 and 1''' ) lowercase_ : Union[str, Any] = state_dict.get('''min_decay''' , self.min_decay ) if not isinstance(self.min_decay , A ): raise ValueError('''Invalid min_decay''' ) lowercase_ : Any = state_dict.get('''optimization_step''' , self.optimization_step ) if not isinstance(self.optimization_step , A ): raise ValueError('''Invalid optimization_step''' ) lowercase_ : Union[str, Any] = state_dict.get('''update_after_step''' , self.update_after_step ) if not isinstance(self.update_after_step , A ): raise ValueError('''Invalid update_after_step''' ) lowercase_ : List[str] = state_dict.get('''use_ema_warmup''' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , A ): raise ValueError('''Invalid use_ema_warmup''' ) lowercase_ : Union[str, Any] = state_dict.get('''inv_gamma''' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('''Invalid inv_gamma''' ) lowercase_ : Any = state_dict.get('''power''' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('''Invalid power''' ) lowercase_ : Optional[int] = state_dict.get('''shadow_params''' , A ) if shadow_params is not None: lowercase_ : Any = shadow_params if not isinstance(self.shadow_params , A ): raise ValueError('''shadow_params must be a list''' ) if not all(isinstance(A , torch.Tensor ) for p in self.shadow_params ): raise ValueError('''shadow_params must all be Tensors''' )
33
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _UpperCAmelCase ( TensorFormatter[Mapping, "torch.Tensor", Mapping] ): def __init__( self : Any , A : int=None , **A : str ) -> Union[str, Any]: super().__init__(features=A ) lowercase_ : Union[str, Any] = torch_tensor_kwargs import torch # noqa import torch at initialization def A ( self : Dict , A : int ) -> List[Any]: import torch if isinstance(A , A ) and column: if all( isinstance(A , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(A ) return column def A ( self : int , A : Any ) -> Optional[Any]: import torch if isinstance(A , (str, bytes, type(A )) ): return value elif isinstance(A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() lowercase_ : Any = {} if isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): lowercase_ : Any = {'''dtype''': torch.intaa} elif isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): lowercase_ : Dict = {'''dtype''': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A , PIL.Image.Image ): lowercase_ : Dict = np.asarray(A ) return torch.tensor(A , **{**default_dtype, **self.torch_tensor_kwargs} ) def A ( self : Union[str, Any] , A : Optional[int] ) -> str: import torch # support for torch, tf, jax etc. if hasattr(A , '''__array__''' ) and not isinstance(A , torch.Tensor ): lowercase_ : Optional[int] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) elif isinstance(A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) return self._tensorize(A ) def A ( self : Dict , A : dict ) -> Tuple: return map_nested(self._recursive_tensorize , A , map_list=A ) def A ( self : str , A : pa.Table ) -> Mapping: lowercase_ : Optional[Any] = self.numpy_arrow_extractor().extract_row(A ) lowercase_ : str = self.python_features_decoder.decode_row(A ) return self.recursive_tensorize(A ) def A ( self : List[Any] , A : pa.Table ) -> "torch.Tensor": lowercase_ : List[str] = self.numpy_arrow_extractor().extract_column(A ) lowercase_ : str = self.python_features_decoder.decode_column(A , pa_table.column_names[0] ) lowercase_ : Optional[int] = self.recursive_tensorize(A ) lowercase_ : Any = self._consolidate(A ) return column def A ( self : List[str] , A : pa.Table ) -> Mapping: lowercase_ : Optional[int] = self.numpy_arrow_extractor().extract_batch(A ) lowercase_ : int = self.python_features_decoder.decode_batch(A ) lowercase_ : Dict = self.recursive_tensorize(A ) for column_name in batch: lowercase_ : Optional[Any] = self._consolidate(batch[column_name] ) return batch
33
1
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
33
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __A : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __A : Union[str, Any] = logging.get_logger(__name__) # General docstring __A : Tuple = '''MobileNetV1Config''' # Base docstring __A : Union[str, Any] = '''google/mobilenet_v1_1.0_224''' __A : Union[str, Any] = [1, 1_024, 7, 7] # Image classification docstring __A : Optional[Any] = '''google/mobilenet_v1_1.0_224''' __A : List[Any] = '''tabby, tabby cat''' __A : Union[str, Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=None ): lowercase_ : str = {} if isinstance(__snake_case , __snake_case ): lowercase_ : Union[str, Any] = model.mobilenet_va else: lowercase_ : Optional[Any] = model lowercase_ : Union[str, Any] = '''MobilenetV1/Conv2d_0/''' lowercase_ : Union[str, Any] = backbone.conv_stem.convolution.weight lowercase_ : Optional[Any] = backbone.conv_stem.normalization.bias lowercase_ : Union[str, Any] = backbone.conv_stem.normalization.weight lowercase_ : Any = backbone.conv_stem.normalization.running_mean lowercase_ : int = backbone.conv_stem.normalization.running_var for i in range(1_3 ): lowercase_ : Optional[int] = i + 1 lowercase_ : Union[str, Any] = i * 2 lowercase_ : Optional[Any] = backbone.layer[pt_index] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' lowercase_ : str = pointer.convolution.weight lowercase_ : int = pointer.normalization.bias lowercase_ : Any = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Union[str, Any] = pointer.normalization.running_var lowercase_ : Any = backbone.layer[pt_index + 1] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' lowercase_ : int = pointer.convolution.weight lowercase_ : str = pointer.normalization.bias lowercase_ : Tuple = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Any = pointer.normalization.running_var if isinstance(__snake_case , __snake_case ): lowercase_ : Optional[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase_ : Any = model.classifier.weight lowercase_ : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase_ : Tuple = tf.train.list_variables(__snake_case ) lowercase_ : int = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) lowercase_ : Optional[Any] = tf.train.load_variable(__snake_case , __snake_case ) lowercase_ : Optional[int] = array # Build TF to PyTorch weights loading map lowercase_ : Any = _build_tf_to_pytorch_map(__snake_case , __snake_case , __snake_case ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue lowercase_ : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase_ : Any = np.transpose(__snake_case , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase_ : Optional[int] = array.squeeze().transpose() else: lowercase_ : Optional[int] = np.transpose(__snake_case , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) lowercase_ : str = torch.from_numpy(__snake_case ) tf_weights.pop(__snake_case , __snake_case ) tf_weights.pop(name + '''/RMSProp''' , __snake_case ) tf_weights.pop(name + '''/RMSProp_1''' , __snake_case ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __snake_case ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def lowercase ( __snake_case : torch.Tensor , __snake_case : nn.Convad ): lowercase_ , lowercase_ : Optional[int] = features.shape[-2:] lowercase_ , lowercase_ : str = conv_layer.stride lowercase_ , lowercase_ : Tuple = conv_layer.kernel_size if in_height % stride_height == 0: lowercase_ : Dict = max(kernel_height - stride_height , 0 ) else: lowercase_ : List[Any] = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase_ : str = max(kernel_width - stride_width , 0 ) else: lowercase_ : int = max(kernel_width - (in_width % stride_width) , 0 ) lowercase_ : int = pad_along_width // 2 lowercase_ : Union[str, Any] = pad_along_width - pad_left lowercase_ : Tuple = pad_along_height // 2 lowercase_ : List[str] = pad_along_height - pad_top lowercase_ : str = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__snake_case , __snake_case , '''constant''' , 0.0 ) class _UpperCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : MobileNetVaConfig , A : int , A : int , A : int , A : Optional[int] = 1 , A : Optional[int] = 1 , A : bool = False , A : Optional[bool] = True , A : Optional[bool or str] = True , ) -> None: super().__init__() lowercase_ : int = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) lowercase_ : Tuple = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase_ : int = nn.Convad( in_channels=A , out_channels=A , kernel_size=A , stride=A , padding=A , groups=A , bias=A , padding_mode='''zeros''' , ) if use_normalization: lowercase_ : Optional[Any] = nn.BatchNormad( num_features=A , eps=config.layer_norm_eps , momentum=0.9997 , affine=A , track_running_stats=A , ) else: lowercase_ : Union[str, Any] = None if use_activation: if isinstance(A , A ): lowercase_ : str = ACTaFN[use_activation] elif isinstance(config.hidden_act , A ): lowercase_ : Any = ACTaFN[config.hidden_act] else: lowercase_ : Tuple = config.hidden_act else: lowercase_ : Tuple = None def A ( self : str , A : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: lowercase_ : List[Any] = apply_tf_padding(A , self.convolution ) lowercase_ : Optional[int] = self.convolution(A ) if self.normalization is not None: lowercase_ : Union[str, Any] = self.normalization(A ) if self.activation is not None: lowercase_ : Optional[int] = self.activation(A ) return features class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = MobileNetVaConfig SCREAMING_SNAKE_CASE_ : int = load_tf_weights_in_mobilenet_va SCREAMING_SNAKE_CASE_ : Optional[Any] = "mobilenet_v1" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any , A : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __A : Union[str, Any] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : str , A : MobileNetVaConfig , A : bool = True ) -> int: super().__init__(A ) lowercase_ : Union[str, Any] = config lowercase_ : List[str] = 32 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase_ : Union[str, Any] = MobileNetVaConvLayer( A , in_channels=config.num_channels , out_channels=A , kernel_size=3 , stride=2 , ) lowercase_ : Optional[Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase_ : List[Any] = nn.ModuleList() for i in range(13 ): lowercase_ : Dict = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=3 , stride=strides[i] , groups=A , ) ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=1 , ) ) lowercase_ : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def A ( self : Any , A : Optional[Any] ) -> Optional[int]: raise NotImplementedError @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : List[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase_ : List[str] = self.conv_stem(A ) lowercase_ : Dict = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase_ : Optional[int] = layer_module(A ) if output_hidden_states: lowercase_ : str = all_hidden_states + (hidden_states,) lowercase_ : Tuple = hidden_states if self.pooler is not None: lowercase_ : Dict = torch.flatten(self.pooler(A ) , start_dim=1 ) else: lowercase_ : Optional[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=A , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : List[str] , A : MobileNetVaConfig ) -> None: super().__init__(A ) lowercase_ : int = config.num_labels lowercase_ : List[str] = MobileNetVaModel(A ) lowercase_ : Union[str, Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase_ : Tuple = nn.Dropout(config.classifier_dropout_prob , inplace=A ) lowercase_ : int = nn.Linear(A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[Any] = self.mobilenet_va(A , output_hidden_states=A , return_dict=A ) lowercase_ : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Dict = self.classifier(self.dropout(A ) ) lowercase_ : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : Optional[Any] = '''single_label_classification''' else: lowercase_ : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : str = MSELoss() if self.num_labels == 1: lowercase_ : List[str] = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(A , A ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[Any] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : str = BCEWithLogitsLoss() lowercase_ : List[Any] = loss_fct(A , A ) if not return_dict: lowercase_ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A , logits=A , hidden_states=outputs.hidden_states , )
33
1
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __A : Union[str, Any] = logging.get_logger(__name__) # General docstring __A : Tuple = '''MobileNetV1Config''' # Base docstring __A : Union[str, Any] = '''google/mobilenet_v1_1.0_224''' __A : Union[str, Any] = [1, 1_024, 7, 7] # Image classification docstring __A : Optional[Any] = '''google/mobilenet_v1_1.0_224''' __A : List[Any] = '''tabby, tabby cat''' __A : Union[str, Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=None ): lowercase_ : str = {} if isinstance(__snake_case , __snake_case ): lowercase_ : Union[str, Any] = model.mobilenet_va else: lowercase_ : Optional[Any] = model lowercase_ : Union[str, Any] = '''MobilenetV1/Conv2d_0/''' lowercase_ : Union[str, Any] = backbone.conv_stem.convolution.weight lowercase_ : Optional[Any] = backbone.conv_stem.normalization.bias lowercase_ : Union[str, Any] = backbone.conv_stem.normalization.weight lowercase_ : Any = backbone.conv_stem.normalization.running_mean lowercase_ : int = backbone.conv_stem.normalization.running_var for i in range(1_3 ): lowercase_ : Optional[int] = i + 1 lowercase_ : Union[str, Any] = i * 2 lowercase_ : Optional[Any] = backbone.layer[pt_index] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' lowercase_ : str = pointer.convolution.weight lowercase_ : int = pointer.normalization.bias lowercase_ : Any = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Union[str, Any] = pointer.normalization.running_var lowercase_ : Any = backbone.layer[pt_index + 1] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' lowercase_ : int = pointer.convolution.weight lowercase_ : str = pointer.normalization.bias lowercase_ : Tuple = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Any = pointer.normalization.running_var if isinstance(__snake_case , __snake_case ): lowercase_ : Optional[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase_ : Any = model.classifier.weight lowercase_ : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase_ : Tuple = tf.train.list_variables(__snake_case ) lowercase_ : int = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) lowercase_ : Optional[Any] = tf.train.load_variable(__snake_case , __snake_case ) lowercase_ : Optional[int] = array # Build TF to PyTorch weights loading map lowercase_ : Any = _build_tf_to_pytorch_map(__snake_case , __snake_case , __snake_case ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue lowercase_ : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase_ : Any = np.transpose(__snake_case , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase_ : Optional[int] = array.squeeze().transpose() else: lowercase_ : Optional[int] = np.transpose(__snake_case , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) lowercase_ : str = torch.from_numpy(__snake_case ) tf_weights.pop(__snake_case , __snake_case ) tf_weights.pop(name + '''/RMSProp''' , __snake_case ) tf_weights.pop(name + '''/RMSProp_1''' , __snake_case ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __snake_case ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def lowercase ( __snake_case : torch.Tensor , __snake_case : nn.Convad ): lowercase_ , lowercase_ : Optional[int] = features.shape[-2:] lowercase_ , lowercase_ : str = conv_layer.stride lowercase_ , lowercase_ : Tuple = conv_layer.kernel_size if in_height % stride_height == 0: lowercase_ : Dict = max(kernel_height - stride_height , 0 ) else: lowercase_ : List[Any] = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase_ : str = max(kernel_width - stride_width , 0 ) else: lowercase_ : int = max(kernel_width - (in_width % stride_width) , 0 ) lowercase_ : int = pad_along_width // 2 lowercase_ : Union[str, Any] = pad_along_width - pad_left lowercase_ : Tuple = pad_along_height // 2 lowercase_ : List[str] = pad_along_height - pad_top lowercase_ : str = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__snake_case , __snake_case , '''constant''' , 0.0 ) class _UpperCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : MobileNetVaConfig , A : int , A : int , A : int , A : Optional[int] = 1 , A : Optional[int] = 1 , A : bool = False , A : Optional[bool] = True , A : Optional[bool or str] = True , ) -> None: super().__init__() lowercase_ : int = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) lowercase_ : Tuple = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase_ : int = nn.Convad( in_channels=A , out_channels=A , kernel_size=A , stride=A , padding=A , groups=A , bias=A , padding_mode='''zeros''' , ) if use_normalization: lowercase_ : Optional[Any] = nn.BatchNormad( num_features=A , eps=config.layer_norm_eps , momentum=0.9997 , affine=A , track_running_stats=A , ) else: lowercase_ : Union[str, Any] = None if use_activation: if isinstance(A , A ): lowercase_ : str = ACTaFN[use_activation] elif isinstance(config.hidden_act , A ): lowercase_ : Any = ACTaFN[config.hidden_act] else: lowercase_ : Tuple = config.hidden_act else: lowercase_ : Tuple = None def A ( self : str , A : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: lowercase_ : List[Any] = apply_tf_padding(A , self.convolution ) lowercase_ : Optional[int] = self.convolution(A ) if self.normalization is not None: lowercase_ : Union[str, Any] = self.normalization(A ) if self.activation is not None: lowercase_ : Optional[int] = self.activation(A ) return features class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = MobileNetVaConfig SCREAMING_SNAKE_CASE_ : int = load_tf_weights_in_mobilenet_va SCREAMING_SNAKE_CASE_ : Optional[Any] = "mobilenet_v1" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any , A : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __A : Union[str, Any] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : str , A : MobileNetVaConfig , A : bool = True ) -> int: super().__init__(A ) lowercase_ : Union[str, Any] = config lowercase_ : List[str] = 32 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase_ : Union[str, Any] = MobileNetVaConvLayer( A , in_channels=config.num_channels , out_channels=A , kernel_size=3 , stride=2 , ) lowercase_ : Optional[Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase_ : List[Any] = nn.ModuleList() for i in range(13 ): lowercase_ : Dict = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=3 , stride=strides[i] , groups=A , ) ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=1 , ) ) lowercase_ : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def A ( self : Any , A : Optional[Any] ) -> Optional[int]: raise NotImplementedError @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : List[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase_ : List[str] = self.conv_stem(A ) lowercase_ : Dict = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase_ : Optional[int] = layer_module(A ) if output_hidden_states: lowercase_ : str = all_hidden_states + (hidden_states,) lowercase_ : Tuple = hidden_states if self.pooler is not None: lowercase_ : Dict = torch.flatten(self.pooler(A ) , start_dim=1 ) else: lowercase_ : Optional[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=A , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : List[str] , A : MobileNetVaConfig ) -> None: super().__init__(A ) lowercase_ : int = config.num_labels lowercase_ : List[str] = MobileNetVaModel(A ) lowercase_ : Union[str, Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase_ : Tuple = nn.Dropout(config.classifier_dropout_prob , inplace=A ) lowercase_ : int = nn.Linear(A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[Any] = self.mobilenet_va(A , output_hidden_states=A , return_dict=A ) lowercase_ : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Dict = self.classifier(self.dropout(A ) ) lowercase_ : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : Optional[Any] = '''single_label_classification''' else: lowercase_ : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : str = MSELoss() if self.num_labels == 1: lowercase_ : List[str] = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(A , A ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[Any] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : str = BCEWithLogitsLoss() lowercase_ : List[Any] = loss_fct(A , A ) if not return_dict: lowercase_ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A , logits=A , hidden_states=outputs.hidden_states , )
33
"""simple docstring""" def lowercase ( __snake_case : list[int] ): lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): for j in range(i + 1 , __snake_case ): if numbers[j] < numbers[i]: lowercase_ , lowercase_ : Optional[int] = numbers[j], numbers[i] return numbers if __name__ == "__main__": __A : int = input('''Enter numbers separated by a comma:\n''').strip() __A : Any = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
33
1
"""simple docstring""" import math import sys import cva import numpy as np def lowercase ( __snake_case : np.ndarray , __snake_case : float ): # For applying gaussian function for each element in matrix. lowercase_ : Union[str, Any] = math.sqrt(__snake_case ) lowercase_ : Any = 1 / (sigma * math.sqrt(2 * math.pi )) return cons * np.exp(-((img / sigma) ** 2) * 0.5 ) def lowercase ( __snake_case : np.ndarray , __snake_case : int , __snake_case : int , __snake_case : int ): lowercase_ : List[str] = kernel_size // 2 return img[x - half : x + half + 1, y - half : y + half + 1] def lowercase ( __snake_case : int , __snake_case : float ): # Creates a gaussian kernel of given dimension. lowercase_ : Tuple = np.zeros((kernel_size, kernel_size) ) for i in range(0 , __snake_case ): for j in range(0 , __snake_case ): lowercase_ : List[str] = math.sqrt( abs(i - kernel_size // 2 ) ** 2 + abs(j - kernel_size // 2 ) ** 2 ) return vec_gaussian(__snake_case , __snake_case ) def lowercase ( __snake_case : np.ndarray , __snake_case : float , __snake_case : float , __snake_case : int , ): lowercase_ : Tuple = np.zeros(img.shape ) lowercase_ : Union[str, Any] = get_gauss_kernel(__snake_case , __snake_case ) lowercase_ , lowercase_ : List[Any] = img.shape for i in range(kernel_size // 2 , size_x - kernel_size // 2 ): for j in range(kernel_size // 2 , size_y - kernel_size // 2 ): lowercase_ : str = get_slice(__snake_case , __snake_case , __snake_case , __snake_case ) lowercase_ : Any = img_s - img_s[kernel_size // 2, kernel_size // 2] lowercase_ : List[Any] = vec_gaussian(__snake_case , __snake_case ) lowercase_ : List[Any] = np.multiply(__snake_case , __snake_case ) lowercase_ : Union[str, Any] = np.multiply(__snake_case , __snake_case ) lowercase_ : Any = np.sum(__snake_case ) / np.sum(__snake_case ) lowercase_ : Optional[Any] = val return imga def lowercase ( __snake_case : list ): lowercase_ : Optional[Any] = args[1] if args[1:] else '''../image_data/lena.jpg''' lowercase_ : Dict = float(args[2] ) if args[2:] else 1.0 lowercase_ : int = float(args[3] ) if args[3:] else 1.0 if args[4:]: lowercase_ : str = int(args[4] ) lowercase_ : Optional[int] = kernel_size + abs(kernel_size % 2 - 1 ) else: lowercase_ : Dict = 5 return filename, spatial_variance, intensity_variance, kernel_size if __name__ == "__main__": __A , __A , __A , __A : List[str] = parse_args(sys.argv) __A : str = cva.imread(filename, 0) cva.imshow('''input image''', img) __A : str = img / 255 __A : Any = out.astype('''float32''') __A : List[Any] = bilateral_filter(out, spatial_variance, intensity_variance, kernel_size) __A : Any = out * 255 __A : Optional[int] = np.uinta(out) cva.imshow('''output image''', out) cva.waitKey(0) cva.destroyAllWindows()
33
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any] , A : Any , A : Tuple=7 , A : Tuple=3 , A : Optional[Any]=30 , A : List[Any]=4_00 , A : Tuple=True , A : Dict=None , A : List[str]=True , A : Optional[int]=[0.5, 0.5, 0.5] , A : Tuple=[0.5, 0.5, 0.5] , A : List[str]=True , A : List[Any]=1 / 2_55 , A : Union[str, Any]=True , ) -> Tuple: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase_ : Optional[int] = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} lowercase_ : Optional[int] = parent lowercase_ : str = batch_size lowercase_ : Tuple = num_channels lowercase_ : str = min_resolution lowercase_ : Any = max_resolution lowercase_ : str = do_resize lowercase_ : Any = size lowercase_ : Optional[int] = do_normalize lowercase_ : List[str] = image_mean lowercase_ : Optional[Any] = image_std lowercase_ : int = do_rescale lowercase_ : List[str] = rescale_factor lowercase_ : int = do_pad def A ( self : Any ) -> str: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Optional[Any] , A : int , A : int=False ) -> Tuple: if not batched: lowercase_ : Optional[int] = image_inputs[0] if isinstance(A , Image.Image ): lowercase_ , lowercase_ : int = image.size else: lowercase_ , lowercase_ : Tuple = image.shape[1], image.shape[2] if w < h: lowercase_ : int = int(self.size['''shortest_edge'''] * h / w ) lowercase_ : Optional[Any] = self.size['''shortest_edge'''] elif w > h: lowercase_ : Optional[Any] = self.size['''shortest_edge'''] lowercase_ : Optional[int] = int(self.size['''shortest_edge'''] * w / h ) else: lowercase_ : Any = self.size['''shortest_edge'''] lowercase_ : Any = self.size['''shortest_edge'''] else: lowercase_ : Tuple = [] for image in image_inputs: lowercase_ , lowercase_ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowercase_ : Union[str, Any] = max(A , key=lambda A : item[0] )[0] lowercase_ : Optional[Any] = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = YolosImageProcessor if is_vision_available() else None def A ( self : Optional[int] ) -> Optional[int]: lowercase_ : Optional[Any] = YolosImageProcessingTester(self ) @property def A ( self : str ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Optional[int] ) -> List[str]: lowercase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , '''image_mean''' ) ) self.assertTrue(hasattr(A , '''image_std''' ) ) self.assertTrue(hasattr(A , '''do_normalize''' ) ) self.assertTrue(hasattr(A , '''do_resize''' ) ) self.assertTrue(hasattr(A , '''size''' ) ) def A ( self : Dict ) -> Tuple: lowercase_ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowercase_ : Tuple = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A ) def A ( self : Optional[int] ) -> Tuple: pass def A ( self : Tuple ) -> int: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowercase_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ , lowercase_ : Dict = self.image_processor_tester.get_expected_values(A , batched=A ) lowercase_ : str = image_processing(A , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : str ) -> Any: # Initialize image_processing lowercase_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowercase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : int = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Optional[int] = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[Any] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[int]: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowercase_ : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Any = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[str] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[Any]: # Initialize image_processings lowercase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) lowercase_ : Tuple = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowercase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowercase_ : Union[str, Any] = image_processing_a.pad(A , return_tensors='''pt''' ) lowercase_ : List[Any] = image_processing_a(A , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1e-4 ) ) @slow def A ( self : str ) -> List[Any]: # prepare image and target lowercase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowercase_ : List[Any] = json.loads(f.read() ) lowercase_ : Tuple = {'''image_id''': 3_97_69, '''annotations''': target} # encode them lowercase_ : Union[str, Any] = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) lowercase_ : List[Any] = image_processing(images=A , annotations=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Union[str, Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Union[str, Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : Tuple = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : List[str] = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : Any = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify orig_size lowercase_ : List[str] = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : Optional[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) ) @slow def A ( self : List[Any] ) -> Dict: # prepare image, target and masks_path lowercase_ : Tuple = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowercase_ : str = json.loads(f.read() ) lowercase_ : int = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} lowercase_ : List[Any] = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowercase_ : int = YolosImageProcessor(format='''coco_panoptic''' ) lowercase_ : Any = image_processing(images=A , annotations=A , masks_path=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Optional[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Tuple = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : List[Any] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : str = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : List[str] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify masks lowercase_ : Dict = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A ) # verify orig_size lowercase_ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) )
33
1
"""simple docstring""" def lowercase ( __snake_case : int ): lowercase_ : List[Any] = generate_pascal_triangle(__snake_case ) for row_idx in range(__snake_case ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=''' ''' ) else: print(triangle[row_idx][col_idx] , end='''''' ) print() def lowercase ( __snake_case : int ): if not isinstance(__snake_case , __snake_case ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) lowercase_ : list[list[int]] = [] for current_row_idx in range(__snake_case ): lowercase_ : int = populate_current_row(__snake_case , __snake_case ) triangle.append(__snake_case ) return triangle def lowercase ( __snake_case : list[list[int]] , __snake_case : int ): lowercase_ : List[Any] = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 lowercase_ , lowercase_ : Any = 1, 1 for current_col_idx in range(1 , __snake_case ): calculate_current_element( __snake_case , __snake_case , __snake_case , __snake_case ) return current_row def lowercase ( __snake_case : list[list[int]] , __snake_case : list[int] , __snake_case : int , __snake_case : int , ): lowercase_ : Dict = triangle[current_row_idx - 1][current_col_idx - 1] lowercase_ : List[str] = triangle[current_row_idx - 1][current_col_idx] lowercase_ : Dict = above_to_left_elt + above_to_right_elt def lowercase ( __snake_case : int ): if not isinstance(__snake_case , __snake_case ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) lowercase_ : list[list[int]] = [[1]] for row_index in range(1 , __snake_case ): lowercase_ : Any = [0] + result[-1] + [0] lowercase_ : Optional[Any] = row_index + 1 # Calculate the number of distinct elements in a row lowercase_ : int = sum(divmod(__snake_case , 2 ) ) lowercase_ : str = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] lowercase_ : Union[str, Any] = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() lowercase_ : Optional[Any] = row_first_half + row_second_half result.append(__snake_case ) return result def lowercase ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(__snake_case : Callable , __snake_case : int ) -> None: lowercase_ : int = F'''{func.__name__}({value})''' lowercase_ : Optional[Any] = timeit(F'''__main__.{call}''' , setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'''{call:38} -- {timing:.4f} seconds''' ) for value in range(1_5 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(__snake_case , __snake_case ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0 ): lowercase_ : str = 0 lowercase_ : List[Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
33
1
"""simple docstring""" from string import ascii_uppercase __A : List[Any] = {char: i for i, char in enumerate(ascii_uppercase)} __A : Optional[int] = dict(enumerate(ascii_uppercase)) def lowercase ( __snake_case : str , __snake_case : str ): lowercase_ : List[Any] = len(__snake_case ) lowercase_ : Dict = 0 while True: if x == i: lowercase_ : Optional[Any] = 0 if len(__snake_case ) == len(__snake_case ): break key += key[i] i += 1 return key def lowercase ( __snake_case : str , __snake_case : str ): lowercase_ : Dict = '''''' lowercase_ : List[str] = 0 for letter in message: if letter == " ": cipher_text += " " else: lowercase_ : Dict = (dicta[letter] - dicta[key_new[i]]) % 2_6 i += 1 cipher_text += dicta[x] return cipher_text def lowercase ( __snake_case : str , __snake_case : str ): lowercase_ : Dict = '''''' lowercase_ : Optional[int] = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: lowercase_ : int = (dicta[letter] + dicta[key_new[i]] + 2_6) % 2_6 i += 1 or_txt += dicta[x] return or_txt def lowercase ( ): lowercase_ : Dict = '''THE GERMAN ATTACK''' lowercase_ : Optional[Any] = '''SECRET''' lowercase_ : int = generate_key(__snake_case , __snake_case ) lowercase_ : List[Any] = cipher_text(__snake_case , __snake_case ) print(F'''Encrypted Text = {s}''' ) print(F'''Original Text = {original_text(__snake_case , __snake_case )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
33
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
1
"""simple docstring""" import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def lowercase ( __snake_case : List[Any] , __snake_case : Any ): assert isinstance(__snake_case , __snake_case ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowercase ( __snake_case : Optional[Any] , __snake_case : int , __snake_case : str , __snake_case : Any ): lowercase_ : str = tmp_path / '''cache''' lowercase_ : str = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase_ : List[Any] = SqlDatasetReader( '''dataset''' , '''sqlite:///''' + sqlite_path , cache_dir=__snake_case , keep_in_memory=__snake_case ).read() _check_sql_dataset(__snake_case , __snake_case ) @require_sqlalchemy @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowercase ( __snake_case : int , __snake_case : int , __snake_case : str , __snake_case : Union[str, Any] ): lowercase_ : List[str] = tmp_path / '''cache''' lowercase_ : Union[str, Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} lowercase_ : Optional[Any] = features.copy() if features else default_expected_features lowercase_ : int = ( Features({feature: Value(__snake_case ) for feature, dtype in features.items()} ) if features is not None else None ) lowercase_ : Any = SqlDatasetReader('''dataset''' , '''sqlite:///''' + sqlite_path , features=__snake_case , cache_dir=__snake_case ).read() _check_sql_dataset(__snake_case , __snake_case ) def lowercase ( __snake_case : List[str] ): with contextlib.closing(sqlitea.connect(__snake_case ) ) as con: lowercase_ : Tuple = con.cursor() cur.execute('''SELECT * FROM dataset''' ) for row in cur: yield row @require_sqlalchemy def lowercase ( __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : List[Any] ): lowercase_ : Optional[Any] = tmp_path / '''cache''' lowercase_ : Union[str, Any] = os.path.join(__snake_case , '''tmp.sql''' ) lowercase_ : Any = SqlDatasetReader('''dataset''' , '''sqlite:///''' + sqlite_path , cache_dir=__snake_case ).read() SqlDatasetWriter(__snake_case , '''dataset''' , '''sqlite:///''' + output_sqlite_path , num_proc=1 ).write() lowercase_ : str = iter_sql_file(__snake_case ) lowercase_ : List[str] = iter_sql_file(__snake_case ) for rowa, rowa in zip(__snake_case , __snake_case ): assert rowa == rowa @require_sqlalchemy def lowercase ( __snake_case : int , __snake_case : Optional[int] , __snake_case : Tuple ): lowercase_ : Any = tmp_path / '''cache''' lowercase_ : List[str] = os.path.join(__snake_case , '''tmp.sql''' ) lowercase_ : Union[str, Any] = SqlDatasetReader('''dataset''' , '''sqlite:///''' + sqlite_path , cache_dir=__snake_case ).read() SqlDatasetWriter(__snake_case , '''dataset''' , '''sqlite:///''' + output_sqlite_path , num_proc=2 ).write() lowercase_ : Dict = iter_sql_file(__snake_case ) lowercase_ : Dict = iter_sql_file(__snake_case ) for rowa, rowa in zip(__snake_case , __snake_case ): assert rowa == rowa @require_sqlalchemy def lowercase ( __snake_case : int , __snake_case : List[str] , __snake_case : str ): lowercase_ : List[Any] = tmp_path / '''cache''' lowercase_ : Union[str, Any] = os.path.join(__snake_case , '''tmp.sql''' ) lowercase_ : Optional[Any] = SqlDatasetReader('''dataset''' , '''sqlite:///''' + sqlite_path , cache_dir=__snake_case ).read() with pytest.raises(__snake_case ): SqlDatasetWriter(__snake_case , '''dataset''' , '''sqlite:///''' + output_sqlite_path , num_proc=0 ).write()
33
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0_0 ): lowercase_ , lowercase_ : str = 1, 1 lowercase_ : List[str] = 2 while True: lowercase_ : Tuple = 0 lowercase_ : List[Any] = fa + fa lowercase_ , lowercase_ : Optional[int] = fa, f index += 1 for _ in str(__snake_case ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class _UpperCAmelCase : def __init__( self : List[Any] , A : Dict , ) -> List[Any]: lowercase_ : List[Any] = parent lowercase_ : str = 13 lowercase_ : str = 7 lowercase_ : Optional[int] = True lowercase_ : List[Any] = True lowercase_ : List[str] = True lowercase_ : Tuple = 99 lowercase_ : str = 32 lowercase_ : Tuple = 2 lowercase_ : int = 4 lowercase_ : Union[str, Any] = 37 lowercase_ : Optional[Any] = '''gelu''' lowercase_ : Any = 0.1 lowercase_ : Dict = 0.1 lowercase_ : int = 5_12 lowercase_ : Optional[Any] = 16 lowercase_ : Any = 2 lowercase_ : Optional[int] = 0.02 lowercase_ : Optional[Any] = 3 lowercase_ : List[Any] = 4 lowercase_ : Union[str, Any] = None def A ( self : Any ) -> Optional[Any]: lowercase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase_ : int = None if self.use_input_mask: lowercase_ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : List[str] = None lowercase_ : Union[str, Any] = None lowercase_ : Dict = None if self.use_labels: lowercase_ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase_ : str = ids_tensor([self.batch_size] , self.num_choices ) lowercase_ : List[Any] = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> List[Any]: ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Dict = self.prepare_config_and_inputs() lowercase_ : Optional[Any] = True lowercase_ : Union[str, Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowercase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : str , A : List[Any] , A : Tuple , A : Tuple , A : Optional[int] , A : Union[str, Any] , A : Any ) -> List[Any]: lowercase_ : Dict = TFEsmModel(config=A ) lowercase_ : Optional[int] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} lowercase_ : Tuple = model(A ) lowercase_ : Any = [input_ids, input_mask] lowercase_ : Any = model(A ) lowercase_ : Optional[int] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Dict , A : List[str] , A : Union[str, Any] , A : Tuple , A : Optional[Any] , A : List[Any] , A : str , A : Dict , A : str , ) -> int: lowercase_ : List[Any] = True lowercase_ : Dict = TFEsmModel(config=A ) lowercase_ : Dict = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } lowercase_ : str = model(A ) lowercase_ : Union[str, Any] = [input_ids, input_mask] lowercase_ : str = model(A , encoder_hidden_states=A ) # Also check the case where encoder outputs are not passed lowercase_ : Dict = model(A , attention_mask=A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , A : Optional[Any] , A : List[str] , A : List[str] , A : Optional[int] , A : List[Any] , A : Tuple ) -> Dict: lowercase_ : Tuple = TFEsmForMaskedLM(config=A ) lowercase_ : Union[str, Any] = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[str] , A : Tuple , A : List[Any] , A : Optional[int] , A : Any , A : Optional[Any] , A : Optional[Any] ) -> int: lowercase_ : Tuple = self.num_labels lowercase_ : List[str] = TFEsmForTokenClassification(config=A ) lowercase_ : Optional[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} lowercase_ : List[str] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Optional[Any] ) -> List[str]: lowercase_ : Dict = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : str = config_and_inputs lowercase_ : Optional[int] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Dict = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : int = ( { "feature-extraction": TFEsmModel, "fill-mask": TFEsmForMaskedLM, "text-classification": TFEsmForSequenceClassification, "token-classification": TFEsmForTokenClassification, "zero-shot": TFEsmForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Dict = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = False def A ( self : int ) -> Tuple: lowercase_ : Any = TFEsmModelTester(self ) lowercase_ : Optional[Any] = ConfigTester(self , config_class=A , hidden_size=37 ) def A ( self : Optional[Any] ) -> Dict: self.config_tester.run_common_tests() def A ( self : Optional[int] ) -> List[str]: lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : Tuple ) -> Any: lowercase_ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*A ) def A ( self : Dict ) -> Optional[Any]: lowercase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A ) def A ( self : List[str] ) -> List[str]: lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) @slow def A ( self : str ) -> Optional[Any]: for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : str = TFEsmModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip('''Protein models do not support embedding resizing.''' ) def A ( self : Optional[Any] ) -> Optional[int]: pass @unittest.skip('''Protein models do not support embedding resizing.''' ) def A ( self : Dict ) -> str: pass def A ( self : str ) -> Optional[Any]: lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : Optional[int] = model_class(A ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer lowercase_ : Optional[Any] = model.get_bias() assert isinstance(A , A ) for k, v in name.items(): assert isinstance(A , tf.Variable ) else: lowercase_ : List[Any] = model.get_output_embeddings() assert x is None lowercase_ : Dict = model.get_bias() assert name is None @require_tf class _UpperCAmelCase ( unittest.TestCase ): @slow def A ( self : Optional[Any] ) -> List[str]: lowercase_ : int = TFEsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) lowercase_ : Optional[int] = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase_ : Union[str, Any] = model(A )[0] lowercase_ : int = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , A ) # compare the actual values for a slice. lowercase_ : str = tf.constant( [ [ [8.921518, -10.589814, -6.4671307], [-6.3967156, -13.911377, -1.1211915], [-7.781247, -13.951557, -3.740592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-2 ) ) @slow def A ( self : Dict ) -> List[str]: lowercase_ : Optional[int] = TFEsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) lowercase_ : Optional[Any] = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) lowercase_ : Any = model(A )[0] # compare the actual values for a slice. lowercase_ : List[Any] = tf.constant( [ [ [0.14443092, 0.54125327, 0.3247739], [0.30340484, 0.00526676, 0.31077722], [0.32278043, -0.24987096, 0.3414628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : Union[str, Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = "vit_mae" def __init__( self : Dict , A : List[str]=7_68 , A : Any=12 , A : Union[str, Any]=12 , A : Tuple=30_72 , A : Any="gelu" , A : Tuple=0.0 , A : List[str]=0.0 , A : Tuple=0.02 , A : Tuple=1e-12 , A : int=2_24 , A : Dict=16 , A : int=3 , A : Tuple=True , A : Tuple=16 , A : Optional[Any]=5_12 , A : Union[str, Any]=8 , A : List[Any]=20_48 , A : Dict=0.75 , A : Any=False , **A : Optional[int] , ) -> Union[str, Any]: super().__init__(**A ) lowercase_ : List[Any] = hidden_size lowercase_ : str = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : List[Any] = hidden_dropout_prob lowercase_ : int = attention_probs_dropout_prob lowercase_ : int = initializer_range lowercase_ : Dict = layer_norm_eps lowercase_ : Optional[Any] = image_size lowercase_ : str = patch_size lowercase_ : Dict = num_channels lowercase_ : Any = qkv_bias lowercase_ : Union[str, Any] = decoder_num_attention_heads lowercase_ : Optional[Any] = decoder_hidden_size lowercase_ : List[str] = decoder_num_hidden_layers lowercase_ : List[Any] = decoder_intermediate_size lowercase_ : Optional[Any] = mask_ratio lowercase_ : Optional[Any] = norm_pix_loss
33
1
"""simple docstring""" def lowercase ( __snake_case : int ): if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: lowercase_ : Dict = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase ( __snake_case : int ): lowercase_ : str = 0 lowercase_ : List[str] = 2 while digits < n: index += 1 lowercase_ : Any = len(str(fibonacci(__snake_case ) ) ) return index def lowercase ( __snake_case : int = 1_0_0_0 ): return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
"""simple docstring""" def lowercase ( __snake_case : int ): if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: lowercase_ : Dict = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase ( __snake_case : int ): lowercase_ : str = 0 lowercase_ : List[str] = 2 while digits < n: index += 1 lowercase_ : Any = len(str(fibonacci(__snake_case ) ) ) return index def lowercase ( __snake_case : int = 1_0_0_0 ): return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm __A : List[str] = logging.get_logger(__name__) @dataclass class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Any = [ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self : str , **A : List[Any] ) -> Union[str, Any]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: lowercase_ : List[str] = deprecated_arg[3:] setattr(self , A , not kwargs.pop(A ) ) logger.warning( F'''{deprecated_arg} is depreciated. Please use --no_{positive_arg} or''' F''' {positive_arg}={kwargs[positive_arg]}''' ) lowercase_ : Optional[Any] = kwargs.pop('''torchscript''' , self.torchscript ) lowercase_ : Tuple = kwargs.pop('''torch_xla_tpu_print_metrics''' , self.torch_xla_tpu_print_metrics ) lowercase_ : Any = kwargs.pop('''fp16_opt_level''' , self.fpaa_opt_level ) super().__init__(**A ) SCREAMING_SNAKE_CASE_ : bool = field(default=_A , metadata={"help": "Trace the models using torchscript"} ) SCREAMING_SNAKE_CASE_ : bool = field(default=_A , metadata={"help": "Print Xla/PyTorch tpu metrics"} ) SCREAMING_SNAKE_CASE_ : str = field( default="O1" , metadata={ "help": ( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. " "See details at https://nvidia.github.io/apex/amp.html" ) } , ) @cached_property def A ( self : Optional[Any] ) -> Tuple["torch.device", int]: requires_backends(self , ['''torch'''] ) logger.info('''PyTorch: setting up devices''' ) if not self.cuda: lowercase_ : List[Any] = torch.device('''cpu''' ) lowercase_ : Optional[int] = 0 elif is_torch_tpu_available(): lowercase_ : Optional[Any] = xm.xla_device() lowercase_ : Union[str, Any] = 0 else: lowercase_ : Tuple = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) lowercase_ : str = torch.cuda.device_count() return device, n_gpu @property def A ( self : Any ) -> str: return is_torch_tpu_available() and self.tpu @property def A ( self : str ) -> int: requires_backends(self , ['''torch'''] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def A ( self : int ) -> "torch.device": requires_backends(self , ['''torch'''] ) return self._setup_devices[0] @property def A ( self : List[str] ) -> str: requires_backends(self , ['''torch'''] ) return self._setup_devices[1] @property def A ( self : List[str] ) -> Optional[Any]: return self.n_gpu > 0
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __A : List[str] = { '''configuration_mobilenet_v2''': [ '''MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileNetV2Config''', '''MobileNetV2OnnxConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''MobileNetV2FeatureExtractor'''] __A : Optional[int] = ['''MobileNetV2ImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileNetV2ForImageClassification''', '''MobileNetV2ForSemanticSegmentation''', '''MobileNetV2Model''', '''MobileNetV2PreTrainedModel''', '''load_tf_weights_in_mobilenet_v2''', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self : int , A : Tuple , A : int=3 , A : List[str]=32 , A : Dict=3 , A : Any=10 , A : Dict=[10, 20, 30, 40] , A : Optional[Any]=[1, 1, 2, 1] , A : Union[str, Any]=True , A : Optional[Any]=True , A : Any="relu" , A : Optional[Any]=3 , A : Tuple=None , ) -> Dict: lowercase_ : str = parent lowercase_ : List[Any] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : int = num_channels lowercase_ : int = embeddings_size lowercase_ : str = hidden_sizes lowercase_ : List[str] = depths lowercase_ : Dict = is_training lowercase_ : int = use_labels lowercase_ : Any = hidden_act lowercase_ : List[Any] = num_labels lowercase_ : Tuple = scope lowercase_ : Optional[Any] = len(A ) def A ( self : str ) -> Tuple: lowercase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_labels ) lowercase_ : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Dict ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def A ( self : str , A : Tuple , A : str , A : str ) -> str: lowercase_ : str = TFResNetModel(config=A ) lowercase_ : Union[str, Any] = model(A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Any , A : int , A : List[Any] , A : Optional[Any] ) -> Optional[Any]: lowercase_ : Tuple = self.num_labels lowercase_ : Union[str, Any] = TFResNetForImageClassification(A ) lowercase_ : Tuple = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : Dict = config_and_inputs lowercase_ : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Any = False def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : int = TFResNetModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Dict ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Dict ) -> List[Any]: return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def A ( self : Any ) -> Any: pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def A ( self : List[str] ) -> Optional[Any]: pass def A ( self : str ) -> Tuple: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(A ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : str = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : List[str] ) -> Tuple: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : List[Any] ) -> List[str]: def check_hidden_states_output(A : Union[str, Any] , A : int , A : List[Any] ): lowercase_ : int = model_class(A ) lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Any = self.model_tester.num_stages self.assertEqual(len(A ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase_ : List[str] = layer_type lowercase_ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : Optional[Any] = True check_hidden_states_output(A , A , A ) def A ( self : Optional[int] ) -> Tuple: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def A ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFResNetModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase ( ): lowercase_ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Any ) -> Optional[int]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A ( self : Any ) -> Optional[int]: lowercase_ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase_ : List[Any] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : List[str] = image_processor(images=A , return_tensors='''tf''' ) # forward pass lowercase_ : Tuple = model(**A ) # verify the logits lowercase_ : Optional[int] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Optional[Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A , atol=1e-4 ) )
33
"""simple docstring""" from __future__ import annotations __A : List[Any] = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __A : str = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): lowercase_ : float = -1 for j in range(i + 1 , __snake_case ): if arr[i] < arr[j]: lowercase_ : List[str] = arr[j] break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] for i, outer in enumerate(__snake_case ): lowercase_ : float = -1 for inner in arr[i + 1 :]: if outer < inner: lowercase_ : List[Any] = inner break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = len(__snake_case ) lowercase_ : list[float] = [] lowercase_ : list[float] = [-1] * arr_size for index in reversed(range(__snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowercase_ : Optional[Any] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __A : int = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
33
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class _UpperCAmelCase : def __init__( self : List[str] , A : List[str] , A : List[str]=2 , A : Tuple=True , A : Union[str, Any]=False , A : Optional[int]=10 , A : Dict=3 , A : Union[str, Any]=32 * 8 , A : Union[str, Any]=32 * 8 , A : Optional[int]=4 , A : Any=64 , ) -> str: lowercase_ : List[str] = parent lowercase_ : Any = batch_size lowercase_ : Union[str, Any] = is_training lowercase_ : Optional[Any] = use_auxiliary_loss lowercase_ : List[Any] = num_queries lowercase_ : Any = num_channels lowercase_ : Tuple = min_size lowercase_ : List[str] = max_size lowercase_ : str = num_labels lowercase_ : Tuple = hidden_dim lowercase_ : int = hidden_dim def A ( self : int ) -> Dict: lowercase_ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( A ) lowercase_ : Dict = torch.ones([self.batch_size, self.min_size, self.max_size] , device=A ) lowercase_ : Tuple = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=A ) > 0.5 ).float() lowercase_ : Optional[int] = (torch.rand((self.batch_size, self.num_labels) , device=A ) > 0.5).long() lowercase_ : List[str] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def A ( self : Optional[int] ) -> Optional[Any]: lowercase_ : Optional[Any] = MaskaFormerConfig( hidden_size=self.hidden_dim , ) lowercase_ : str = self.num_queries lowercase_ : Dict = self.num_labels lowercase_ : Union[str, Any] = [1, 1, 1, 1] lowercase_ : Tuple = self.num_channels lowercase_ : int = 64 lowercase_ : Optional[Any] = 1_28 lowercase_ : int = self.hidden_dim lowercase_ : List[Any] = self.hidden_dim lowercase_ : Union[str, Any] = self.hidden_dim return config def A ( self : Optional[int] ) -> Dict: lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : List[Any] = self.prepare_config_and_inputs() lowercase_ : List[str] = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def A ( self : List[Any] , A : int , A : int ) -> Tuple: lowercase_ : List[str] = output.encoder_hidden_states lowercase_ : Tuple = output.pixel_decoder_hidden_states lowercase_ : int = output.transformer_decoder_hidden_states self.parent.assertTrue(len(A ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) , config.decoder_layers ) def A ( self : Any , A : Optional[Any] , A : Optional[Any] , A : int , A : int=False ) -> Union[str, Any]: with torch.no_grad(): lowercase_ : List[Any] = MaskaFormerModel(config=A ) model.to(A ) model.eval() lowercase_ : Any = model(pixel_values=A , pixel_mask=A ) lowercase_ : Tuple = model(A , output_hidden_states=A ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(A , A ) def A ( self : Optional[int] , A : Union[str, Any] , A : Tuple , A : str , A : Optional[Any] , A : Tuple ) -> Optional[Any]: lowercase_ : Union[str, Any] = MaskaFormerForUniversalSegmentation(config=A ) model.to(A ) model.eval() def comm_check_on_output(A : List[Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): lowercase_ : Dict = model(pixel_values=A , pixel_mask=A ) lowercase_ : Dict = model(A ) comm_check_on_output(A ) lowercase_ : str = model( pixel_values=A , pixel_mask=A , mask_labels=A , class_labels=A ) comm_check_on_output(A ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Tuple = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE_ : Optional[int] = {"feature-extraction": MaskaFormerModel} if is_torch_available() else {} SCREAMING_SNAKE_CASE_ : Tuple = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = False SCREAMING_SNAKE_CASE_ : int = False def A ( self : Dict ) -> Dict: lowercase_ : Any = MaskaFormerModelTester(self ) lowercase_ : Union[str, Any] = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Union[str, Any] ) -> List[str]: self.config_tester.run_common_tests() def A ( self : Optional[Any] ) -> Optional[Any]: lowercase_ , lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A , **A , output_hidden_states=A ) def A ( self : List[str] ) -> Any: lowercase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*A ) @unittest.skip(reason='''Mask2Former does not use inputs_embeds''' ) def A ( self : Dict ) -> List[Any]: pass @unittest.skip(reason='''Mask2Former does not have a get_input_embeddings method''' ) def A ( self : Union[str, Any] ) -> Optional[Any]: pass @unittest.skip(reason='''Mask2Former is not a generative model''' ) def A ( self : int ) -> List[Any]: pass @unittest.skip(reason='''Mask2Former does not use token embeddings''' ) def A ( self : Tuple ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip( reason='''Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def A ( self : Dict ) -> List[str]: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def A ( self : Tuple ) -> int: pass def A ( self : Optional[int] ) -> int: lowercase_ , lowercase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : List[str] = model_class(A ) lowercase_ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : Any = [*signature.parameters.keys()] lowercase_ : Optional[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) @slow def A ( self : Tuple ) -> Optional[Any]: for model_name in ["facebook/mask2former-swin-small-coco-instance"]: lowercase_ : Tuple = MaskaFormerModel.from_pretrained(A ) self.assertIsNotNone(A ) def A ( self : Tuple ) -> List[Any]: lowercase_ : Dict = (self.model_tester.min_size,) * 2 lowercase_ : Optional[Any] = { '''pixel_values''': torch.randn((2, 3, *size) , device=A ), '''mask_labels''': torch.randn((2, 10, *size) , device=A ), '''class_labels''': torch.zeros(2 , 10 , device=A ).long(), } lowercase_ : int = self.model_tester.get_config() lowercase_ : str = MaskaFormerForUniversalSegmentation(A ).to(A ) lowercase_ : Optional[int] = model(**A ) self.assertTrue(outputs.loss is not None ) def A ( self : str ) -> List[str]: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A , **A , output_hidden_states=A ) def A ( self : List[Any] ) -> Optional[int]: lowercase_ , lowercase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : List[str] = model_class(A ).to(A ) lowercase_ : Any = model(**A , output_attentions=A ) self.assertTrue(outputs.attentions is not None ) def A ( self : int ) -> str: if not self.model_tester.is_training: return lowercase_ : Any = self.all_model_classes[1] lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() lowercase_ : Tuple = model_class(A ) model.to(A ) model.train() lowercase_ : Optional[int] = model(A , mask_labels=A , class_labels=A ).loss loss.backward() def A ( self : Any ) -> Union[str, Any]: lowercase_ : Optional[int] = self.all_model_classes[1] lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() lowercase_ : int = True lowercase_ : Tuple = True lowercase_ : List[str] = model_class(A ).to(A ) model.train() lowercase_ : Dict = model(A , mask_labels=A , class_labels=A ) lowercase_ : Optional[Any] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() lowercase_ : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() lowercase_ : List[str] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() lowercase_ : Union[str, Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=A ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __A : Any = 1E-4 def lowercase ( ): lowercase_ : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Optional[int] ) -> str: return "facebook/mask2former-swin-small-coco-instance" @cached_property def A ( self : Optional[int] ) -> List[str]: return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def A ( self : Tuple ) -> Tuple: lowercase_ : Optional[Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(A ) lowercase_ : Any = self.default_image_processor lowercase_ : int = prepare_img() lowercase_ : Optional[int] = image_processor(A , return_tensors='''pt''' ).to(A ) lowercase_ : Dict = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A , (1, 3, 3_84, 3_84) ) with torch.no_grad(): lowercase_ : Dict = model(**A ) lowercase_ : List[str] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(A ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , A , atol=A ) ) lowercase_ : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(A ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , A , atol=A ) ) lowercase_ : Tuple = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(A ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , A , atol=A ) ) def A ( self : Union[str, Any] ) -> Dict: lowercase_ : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() lowercase_ : Optional[int] = self.default_image_processor lowercase_ : List[str] = prepare_img() lowercase_ : List[Any] = image_processor(A , return_tensors='''pt''' ).to(A ) lowercase_ : Any = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A , (1, 3, 3_84, 3_84) ) with torch.no_grad(): lowercase_ : List[Any] = model(**A ) # masks_queries_logits lowercase_ : Union[str, Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) lowercase_ : Optional[int] = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] lowercase_ : str = torch.tensor(A ).to(A ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , A , atol=A ) ) # class_queries_logits lowercase_ : List[str] = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) lowercase_ : Optional[Any] = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(A ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , A , atol=A ) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() lowercase_ : List[str] = self.default_image_processor lowercase_ : Union[str, Any] = image_processor( [np.zeros((3, 8_00, 13_33) ), np.zeros((3, 8_00, 13_33) )] , segmentation_maps=[np.zeros((3_84, 3_84) ).astype(np.floataa ), np.zeros((3_84, 3_84) ).astype(np.floataa )] , return_tensors='''pt''' , ) lowercase_ : Optional[int] = inputs['''pixel_values'''].to(A ) lowercase_ : Union[str, Any] = [el.to(A ) for el in inputs['''mask_labels''']] lowercase_ : str = [el.to(A ) for el in inputs['''class_labels''']] with torch.no_grad(): lowercase_ : List[Any] = model(**A ) self.assertTrue(outputs.loss is not None )
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __A : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() __A : Union[str, Any] = logging.get_logger('''transformers.models.encodec''') __A : Dict = { '''quantizer.vq.layers.*._codebook.inited''': '''quantizer.layers.*.codebook.inited''', '''quantizer.vq.layers.*._codebook.cluster_size''': '''quantizer.layers.*.codebook.cluster_size''', '''quantizer.vq.layers.*._codebook.embed''': '''quantizer.layers.*.codebook.embed''', '''quantizer.vq.layers.*._codebook.embed_avg''': '''quantizer.layers.*.codebook.embed_avg''', } __A : str = { '''encoder.model.0.conv.conv''': '''encoder.layers.0.conv''', '''encoder.model.1.block.1.conv.conv''': '''encoder.layers.1.block.1.conv''', '''encoder.model.1.block.3.conv.conv''': '''encoder.layers.1.block.3.conv''', '''encoder.model.1.shortcut.conv.conv''': '''encoder.layers.1.shortcut.conv''', '''encoder.model.3.conv.conv''': '''encoder.layers.3.conv''', '''encoder.model.4.block.1.conv.conv''': '''encoder.layers.4.block.1.conv''', '''encoder.model.4.block.3.conv.conv''': '''encoder.layers.4.block.3.conv''', '''encoder.model.4.shortcut.conv.conv''': '''encoder.layers.4.shortcut.conv''', '''encoder.model.6.conv.conv''': '''encoder.layers.6.conv''', '''encoder.model.7.block.1.conv.conv''': '''encoder.layers.7.block.1.conv''', '''encoder.model.7.block.3.conv.conv''': '''encoder.layers.7.block.3.conv''', '''encoder.model.7.shortcut.conv.conv''': '''encoder.layers.7.shortcut.conv''', '''encoder.model.9.conv.conv''': '''encoder.layers.9.conv''', '''encoder.model.10.block.1.conv.conv''': '''encoder.layers.10.block.1.conv''', '''encoder.model.10.block.3.conv.conv''': '''encoder.layers.10.block.3.conv''', '''encoder.model.10.shortcut.conv.conv''': '''encoder.layers.10.shortcut.conv''', '''encoder.model.12.conv.conv''': '''encoder.layers.12.conv''', '''encoder.model.13.lstm''': '''encoder.layers.13.lstm''', '''encoder.model.15.conv.conv''': '''encoder.layers.15.conv''', } __A : Dict = { '''encoder.model.0.conv.norm''': '''encoder.layers.0.norm''', '''encoder.model.1.block.1.conv.norm''': '''encoder.layers.1.block.1.norm''', '''encoder.model.1.block.3.conv.norm''': '''encoder.layers.1.block.3.norm''', '''encoder.model.1.shortcut.conv.norm''': '''encoder.layers.1.shortcut.norm''', '''encoder.model.3.conv.norm''': '''encoder.layers.3.norm''', '''encoder.model.4.block.1.conv.norm''': '''encoder.layers.4.block.1.norm''', '''encoder.model.4.block.3.conv.norm''': '''encoder.layers.4.block.3.norm''', '''encoder.model.4.shortcut.conv.norm''': '''encoder.layers.4.shortcut.norm''', '''encoder.model.6.conv.norm''': '''encoder.layers.6.norm''', '''encoder.model.7.block.1.conv.norm''': '''encoder.layers.7.block.1.norm''', '''encoder.model.7.block.3.conv.norm''': '''encoder.layers.7.block.3.norm''', '''encoder.model.7.shortcut.conv.norm''': '''encoder.layers.7.shortcut.norm''', '''encoder.model.9.conv.norm''': '''encoder.layers.9.norm''', '''encoder.model.10.block.1.conv.norm''': '''encoder.layers.10.block.1.norm''', '''encoder.model.10.block.3.conv.norm''': '''encoder.layers.10.block.3.norm''', '''encoder.model.10.shortcut.conv.norm''': '''encoder.layers.10.shortcut.norm''', '''encoder.model.12.conv.norm''': '''encoder.layers.12.norm''', '''encoder.model.15.conv.norm''': '''encoder.layers.15.norm''', } __A : int = { '''decoder.model.0.conv.conv''': '''decoder.layers.0.conv''', '''decoder.model.1.lstm''': '''decoder.layers.1.lstm''', '''decoder.model.3.convtr.convtr''': '''decoder.layers.3.conv''', '''decoder.model.4.block.1.conv.conv''': '''decoder.layers.4.block.1.conv''', '''decoder.model.4.block.3.conv.conv''': '''decoder.layers.4.block.3.conv''', '''decoder.model.4.shortcut.conv.conv''': '''decoder.layers.4.shortcut.conv''', '''decoder.model.6.convtr.convtr''': '''decoder.layers.6.conv''', '''decoder.model.7.block.1.conv.conv''': '''decoder.layers.7.block.1.conv''', '''decoder.model.7.block.3.conv.conv''': '''decoder.layers.7.block.3.conv''', '''decoder.model.7.shortcut.conv.conv''': '''decoder.layers.7.shortcut.conv''', '''decoder.model.9.convtr.convtr''': '''decoder.layers.9.conv''', '''decoder.model.10.block.1.conv.conv''': '''decoder.layers.10.block.1.conv''', '''decoder.model.10.block.3.conv.conv''': '''decoder.layers.10.block.3.conv''', '''decoder.model.10.shortcut.conv.conv''': '''decoder.layers.10.shortcut.conv''', '''decoder.model.12.convtr.convtr''': '''decoder.layers.12.conv''', '''decoder.model.13.block.1.conv.conv''': '''decoder.layers.13.block.1.conv''', '''decoder.model.13.block.3.conv.conv''': '''decoder.layers.13.block.3.conv''', '''decoder.model.13.shortcut.conv.conv''': '''decoder.layers.13.shortcut.conv''', '''decoder.model.15.conv.conv''': '''decoder.layers.15.conv''', } __A : Optional[Any] = { '''decoder.model.0.conv.norm''': '''decoder.layers.0.norm''', '''decoder.model.3.convtr.norm''': '''decoder.layers.3.norm''', '''decoder.model.4.block.1.conv.norm''': '''decoder.layers.4.block.1.norm''', '''decoder.model.4.block.3.conv.norm''': '''decoder.layers.4.block.3.norm''', '''decoder.model.4.shortcut.conv.norm''': '''decoder.layers.4.shortcut.norm''', '''decoder.model.6.convtr.norm''': '''decoder.layers.6.norm''', '''decoder.model.7.block.1.conv.norm''': '''decoder.layers.7.block.1.norm''', '''decoder.model.7.block.3.conv.norm''': '''decoder.layers.7.block.3.norm''', '''decoder.model.7.shortcut.conv.norm''': '''decoder.layers.7.shortcut.norm''', '''decoder.model.9.convtr.norm''': '''decoder.layers.9.norm''', '''decoder.model.10.block.1.conv.norm''': '''decoder.layers.10.block.1.norm''', '''decoder.model.10.block.3.conv.norm''': '''decoder.layers.10.block.3.norm''', '''decoder.model.10.shortcut.conv.norm''': '''decoder.layers.10.shortcut.norm''', '''decoder.model.12.convtr.norm''': '''decoder.layers.12.norm''', '''decoder.model.13.block.1.conv.norm''': '''decoder.layers.13.block.1.norm''', '''decoder.model.13.block.3.conv.norm''': '''decoder.layers.13.block.3.norm''', '''decoder.model.13.shortcut.conv.norm''': '''decoder.layers.13.shortcut.norm''', '''decoder.model.15.conv.norm''': '''decoder.layers.15.norm''', } __A : Tuple = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } __A : Optional[int] = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } __A : Optional[int] = [] __A : Dict = [] def lowercase ( __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : List[Any] , __snake_case : str , __snake_case : Tuple ): for attribute in key.split('''.''' ): lowercase_ : Union[str, Any] = getattr(__snake_case , __snake_case ) if weight_type is not None: lowercase_ : Any = getattr(__snake_case , __snake_case ).shape else: lowercase_ : Union[str, Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": lowercase_ : Tuple = value elif weight_type == "weight_g": lowercase_ : Optional[Any] = value elif weight_type == "weight_v": lowercase_ : Tuple = value elif weight_type == "bias": lowercase_ : Optional[int] = value elif weight_type == "running_mean": lowercase_ : Any = value elif weight_type == "running_var": lowercase_ : Optional[int] = value elif weight_type == "num_batches_tracked": lowercase_ : Dict = value elif weight_type == "weight_ih_l0": lowercase_ : Optional[int] = value elif weight_type == "weight_hh_l0": lowercase_ : Tuple = value elif weight_type == "bias_ih_l0": lowercase_ : Optional[int] = value elif weight_type == "bias_hh_l0": lowercase_ : Any = value elif weight_type == "weight_ih_l1": lowercase_ : Optional[int] = value elif weight_type == "weight_hh_l1": lowercase_ : str = value elif weight_type == "bias_ih_l1": lowercase_ : Optional[int] = value elif weight_type == "bias_hh_l1": lowercase_ : str = value else: lowercase_ : List[str] = value logger.info(F'''{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.''' ) def lowercase ( __snake_case : int , __snake_case : Dict ): for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: lowercase_ , lowercase_ : Union[str, Any] = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowercase ( __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Tuple ): lowercase_ : Dict = [] if model_name == "encodec_24khz" or "encodec_32khz": lowercase_ : Dict = MAPPING_24K elif model_name == "encodec_48khz": lowercase_ : Union[str, Any] = MAPPING_48K else: raise ValueError(F'''Unsupported model: {model_name}''' ) for name, value in orig_dict.items(): if should_ignore(__snake_case , __snake_case ): logger.info(F'''{name} was ignored''' ) continue lowercase_ : List[str] = False for key, mapped_key in MAPPING.items(): if "*" in key: lowercase_ , lowercase_ : Dict = key.split('''.*.''' ) if prefix in name and suffix in name: lowercase_ : List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue lowercase_ : Union[str, Any] = True if "*" in mapped_key: lowercase_ : str = name.split(__snake_case )[0].split('''.''' )[-2] lowercase_ : Any = mapped_key.replace('''*''' , __snake_case ) if "weight_g" in name: lowercase_ : Any = '''weight_g''' elif "weight_v" in name: lowercase_ : Optional[int] = '''weight_v''' elif "weight_ih_l0" in name: lowercase_ : int = '''weight_ih_l0''' elif "weight_hh_l0" in name: lowercase_ : int = '''weight_hh_l0''' elif "bias_ih_l0" in name: lowercase_ : Tuple = '''bias_ih_l0''' elif "bias_hh_l0" in name: lowercase_ : Union[str, Any] = '''bias_hh_l0''' elif "weight_ih_l1" in name: lowercase_ : List[str] = '''weight_ih_l1''' elif "weight_hh_l1" in name: lowercase_ : int = '''weight_hh_l1''' elif "bias_ih_l1" in name: lowercase_ : Tuple = '''bias_ih_l1''' elif "bias_hh_l1" in name: lowercase_ : Optional[Any] = '''bias_hh_l1''' elif "bias" in name: lowercase_ : List[Any] = '''bias''' elif "weight" in name: lowercase_ : Dict = '''weight''' elif "running_mean" in name: lowercase_ : List[str] = '''running_mean''' elif "running_var" in name: lowercase_ : Tuple = '''running_var''' elif "num_batches_tracked" in name: lowercase_ : str = '''num_batches_tracked''' else: lowercase_ : int = None set_recursively(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F'''Unused weights: {unused_weights}''' ) @torch.no_grad() def lowercase ( __snake_case : List[str] , __snake_case : Dict , __snake_case : List[Any] , __snake_case : Union[str, Any]=None , __snake_case : Optional[int]=None , ): if config_path is not None: lowercase_ : List[Any] = EncodecConfig.from_pretrained(__snake_case ) else: lowercase_ : Optional[Any] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": lowercase_ : Any = [8, 5, 4, 4] lowercase_ : List[str] = [2.2] lowercase_ : Optional[Any] = 6_4 lowercase_ : int = 3_2_0_0_0 lowercase_ : Union[str, Any] = 2_0_4_8 lowercase_ : Union[str, Any] = False lowercase_ : int = False lowercase_ : str = False elif model_name == "encodec_48khz": lowercase_ : str = [8, 5, 4, 2] lowercase_ : str = [3.0, 6.0, 12.0, 24.0] lowercase_ : str = 4_8_0_0_0 lowercase_ : Optional[Any] = 2 lowercase_ : List[str] = False lowercase_ : int = '''time_group_norm''' lowercase_ : List[str] = True lowercase_ : Dict = 1.0 lowercase_ : Union[str, Any] = 0.01 else: raise ValueError(F'''Unknown model name: {model_name}''' ) lowercase_ : Optional[Any] = EncodecModel(__snake_case ) lowercase_ : Dict = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(__snake_case ) lowercase_ : str = torch.load(__snake_case ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights lowercase_ : List[str] = original_checkpoint['''best_state'''] recursively_load_weights(__snake_case , __snake_case , __snake_case ) model.save_pretrained(__snake_case ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(__snake_case ) model.push_to_hub(__snake_case ) if __name__ == "__main__": __A : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model''', default='''encodec_24khz''', type=str, help='''The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __A : Optional[int] = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
33
"""simple docstring""" def lowercase ( __snake_case : int ): if not isinstance(__snake_case , __snake_case ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
33
1
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __A : List[Any] = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' __A : Dict = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' __A : Any = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): def A ( self : Tuple ) -> List[str]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def A ( self : int , A : Optional[int] , A : List[Any] , A : Any=None , A : Tuple=True , A : List[Any]=False ) -> int: if rouge_types is None: lowercase_ : int = ['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase_ : List[str] = rouge_scorer.RougeScorer(rouge_types=A , use_stemmer=A ) if use_aggregator: lowercase_ : Union[str, Any] = scoring.BootstrapAggregator() else: lowercase_ : List[Any] = [] for ref, pred in zip(A , A ): lowercase_ : Dict = scorer.score(A , A ) if use_aggregator: aggregator.add_scores(A ) else: scores.append(A ) if use_aggregator: lowercase_ : Optional[Any] = aggregator.aggregate() else: lowercase_ : int = {} for key in scores[0]: lowercase_ : Union[str, Any] = [score[key] for score in scores] return result
33
"""simple docstring""" def lowercase ( __snake_case : Optional[int] ): lowercase_ : int = 0 lowercase_ : Optional[Any] = len(__snake_case ) for i in range(n - 1 ): for j in range(i + 1 , __snake_case ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowercase ( __snake_case : str ): if len(__snake_case ) <= 1: return arr, 0 lowercase_ : Optional[Any] = len(__snake_case ) // 2 lowercase_ : List[Any] = arr[0:mid] lowercase_ : Union[str, Any] = arr[mid:] lowercase_ , lowercase_ : Tuple = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = _count_cross_inversions(__snake_case , __snake_case ) lowercase_ : List[Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowercase ( __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Optional[Any] = [] lowercase_ : Any = 0 while i < len(__snake_case ) and j < len(__snake_case ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__snake_case ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__snake_case ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowercase ( ): lowercase_ : Union[str, Any] = [1_0, 2, 1, 5, 5, 2, 1_1] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowercase_ : int = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 8 print('''number of inversions = ''' , __snake_case ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowercase_ : Dict = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) # an empty list should also have zero inversions lowercase_ : List[Any] = [] lowercase_ : Any = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : List[str] = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) if __name__ == "__main__": main()
33
1
"""simple docstring""" from collections import defaultdict def lowercase ( __snake_case : str , __snake_case : str ): lowercase_ : int = first_str.lower().strip() lowercase_ : Any = second_str.lower().strip() # Remove whitespace lowercase_ : int = first_str.replace(''' ''' , '''''' ) lowercase_ : Optional[int] = second_str.replace(''' ''' , '''''' ) # Strings of different lengths are not anagrams if len(__snake_case ) != len(__snake_case ): return False # Default values for count should be 0 lowercase_ : defaultdict[str, int] = defaultdict(__snake_case ) # For each character in input strings, # increment count in the corresponding for i in range(len(__snake_case ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() __A : Optional[Any] = input('''Enter the first string ''').strip() __A : Any = input('''Enter the second string ''').strip() __A : Any = check_anagrams(input_a, input_b) print(F"""{input_a} and {input_b} are {'' if status else 'not '}anagrams.""")
33
"""simple docstring""" __A : Any = { '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
33
1
"""simple docstring""" class _UpperCAmelCase : def __init__( self : List[Any] ) -> Union[str, Any]: lowercase_ : List[str] = {} def A ( self : Dict ) -> None: print(self.vertex ) for i in self.vertex: print(A , ''' -> ''' , ''' -> '''.join([str(A ) for j in self.vertex[i]] ) ) def A ( self : Dict , A : int , A : int ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(A ) else: # else make a new vertex lowercase_ : int = [to_vertex] def A ( self : Any ) -> None: # visited array for storing already visited nodes lowercase_ : int = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(A , A ) def A ( self : Tuple , A : int , A : list ) -> None: # mark start vertex as visited lowercase_ : Optional[Any] = True print(A , end=''' ''' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(A , A ) if __name__ == "__main__": __A : Tuple = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('''DFS:''') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : List[Any] = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): __A : List[str] = True from torch.cuda.amp import autocast __A : str = logging.getLogger(__name__) @dataclass class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default=_A , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) SCREAMING_SNAKE_CASE_ : Optional[bool] = field( default=_A , metadata={"help": "Whether to freeze the feature extractor layers of the model."} ) SCREAMING_SNAKE_CASE_ : Optional[bool] = field( default=_A , metadata={"help": "Whether to log verbose messages or not."} , ) SCREAMING_SNAKE_CASE_ : Optional[float] = field( default=2.0 , metadata={"help": "Maximum temperature for gumbel softmax."} ) SCREAMING_SNAKE_CASE_ : Optional[float] = field( default=0.5 , metadata={"help": "Minimum temperature for gumbel softmax."} ) SCREAMING_SNAKE_CASE_ : Optional[float] = field( default=0.999_995 , metadata={"help": "Decay of gumbel temperature during training."} ) def lowercase ( __snake_case : ModelArguments , __snake_case : TrainingArguments ): logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) lowercase_ : int = logging.WARNING if model_args.verbose_logging: lowercase_ : Any = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): lowercase_ : Tuple = logging.INFO logger.setLevel(__snake_case ) @dataclass class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : str = field( default=_A , metadata={"help": "The name of the dataset to use (via the datasets library)."} ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default=_A , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default="train" , metadata={ "help": "The name of the training data set split to use (via the datasets library). Defaults to 'train'" } , ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default="validation" , metadata={ "help": ( "The name of the validation data set split to use (via the datasets library). Defaults to 'validation'" ) } , ) SCREAMING_SNAKE_CASE_ : Optional[str] = field( default="file" , metadata={"help": "Column in the dataset that contains speech file path. Defaults to 'file'"} , ) SCREAMING_SNAKE_CASE_ : bool = field( default=_A , metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) SCREAMING_SNAKE_CASE_ : Optional[int] = field( default=1 , metadata={ "help": "The percentage of the train set used as validation set in case there's no validation split" } , ) SCREAMING_SNAKE_CASE_ : Optional[int] = field( default=_A , metadata={"help": "The number of processes to use for the preprocessing."} , ) SCREAMING_SNAKE_CASE_ : Optional[float] = field( default=20.0 , metadata={"help": "Filter audio files that are longer than `max_duration_in_seconds` seconds"} ) @dataclass class _UpperCAmelCase : SCREAMING_SNAKE_CASE_ : WavaVecaForPreTraining SCREAMING_SNAKE_CASE_ : WavaVecaFeatureExtractor SCREAMING_SNAKE_CASE_ : Union[bool, str] = "longest" SCREAMING_SNAKE_CASE_ : Optional[int] = None SCREAMING_SNAKE_CASE_ : Optional[int] = None def __call__( self : Union[str, Any] , A : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: # reformat list to dict and set to pytorch format lowercase_ : List[str] = self.feature_extractor.pad( A , max_length=self.max_length , padding=self.padding , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) lowercase_ : Tuple = self.model._get_feat_extract_output_lengths(batch['''input_values'''].shape[-1] ) lowercase_ : Union[str, Any] = batch['''input_values'''].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowercase_ : Any = self.model._get_feat_extract_output_lengths(batch['''attention_mask'''].sum(-1 ) ).to( torch.long ) lowercase_ : str = torch.zeros( (batch_size, mask_indices_seq_length) , dtype=torch.long , device=batch['''input_values'''].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowercase_ : Optional[Any] = 1 lowercase_ : Tuple = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowercase_ : Union[str, Any] = _compute_mask_indices( (batch_size, mask_indices_seq_length) , self.model.config.mask_time_prob , self.model.config.mask_time_length , attention_mask=A , min_masks=2 , ) return batch class _UpperCAmelCase ( _A ): def __init__( self : Optional[Any] , *A : Dict , A : Any=1 , A : Dict=0 , A : Optional[int]=1.0 , **A : str ) -> Tuple: super().__init__(*A , **A ) lowercase_ : Any = 0 lowercase_ : Tuple = max_gumbel_temp lowercase_ : int = min_gumbel_temp lowercase_ : Optional[Any] = gumbel_temp_decay def A ( self : str , A : nn.Module , A : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: model.train() lowercase_ : Any = self._prepare_inputs(A ) if self.use_amp: with autocast(): lowercase_ : Union[str, Any] = self.compute_loss(A , A ) else: lowercase_ : str = self.compute_loss(A , A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowercase_ : Optional[int] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowercase_ : List[Any] = loss.sum() / (inputs['''mask_time_indices''']).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowercase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) return loss.detach() def lowercase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase_ : int = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase_ , lowercase_ , lowercase_ : Optional[Any] = parser.parse_args_into_dataclasses() configure_logger(__snake_case , __snake_case ) # Downloading and loading a dataset from the hub. lowercase_ : Union[str, Any] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowercase_ : Optional[int] = DatasetDict() lowercase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowercase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowercase_ : Dict = DatasetDict() lowercase_ : List[Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split='''validation''' , cache_dir=model_args.cache_dir , ) lowercase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowercase_ : str = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=__snake_case ) def prepare_dataset(__snake_case : int ): # check that all files have the correct sampling rate lowercase_ , lowercase_ : List[str] = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays lowercase_ : Dict = datasets.map( __snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets['''train'''].column_names ) # filter audio files that are too long lowercase_ : Union[str, Any] = vectorized_datasets.filter( lambda __snake_case : len(data['''speech'''] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(__snake_case : Tuple ): return feature_extractor(batch['''speech'''] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` lowercase_ : Union[str, Any] = vectorized_datasets.map( __snake_case , batched=__snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets['''train'''].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowercase_ : str = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( '''PreTraining is only supported for ``config.do_stable_layer_norm=True`` and''' ''' ``config.feat_extract_norm=\'layer\'''' ) lowercase_ : Optional[int] = WavaVecaForPreTraining(__snake_case ) lowercase_ : Union[str, Any] = DataCollatorForWavaVecaPretraining(model=__snake_case , feature_extractor=__snake_case ) lowercase_ : Dict = WavaVecaPreTrainer( model=__snake_case , data_collator=__snake_case , args=__snake_case , train_dataset=vectorized_datasets['''train'''] , eval_dataset=vectorized_datasets['''validation'''] , tokenizer=__snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
33
"""simple docstring""" import argparse import os import re import packaging.version __A : List[str] = '''examples/''' __A : int = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __A : Dict = { '''init''': '''src/transformers/__init__.py''', '''setup''': '''setup.py''', } __A : Optional[int] = '''README.md''' def lowercase ( __snake_case : int , __snake_case : Any , __snake_case : int ): with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : int = f.read() lowercase_ , lowercase_ : List[str] = REPLACE_PATTERNS[pattern] lowercase_ : Union[str, Any] = replace.replace('''VERSION''' , __snake_case ) lowercase_ : Optional[Any] = re_pattern.sub(__snake_case , __snake_case ) with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(__snake_case ) def lowercase ( __snake_case : int ): for folder, directories, fnames in os.walk(__snake_case ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__snake_case , __snake_case ) , __snake_case , pattern='''examples''' ) def lowercase ( __snake_case : Optional[Any] , __snake_case : Optional[Any]=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__snake_case , __snake_case , __snake_case ) if not patch: update_version_in_examples(__snake_case ) def lowercase ( ): lowercase_ : Union[str, Any] = '''🤗 Transformers currently provides the following architectures''' lowercase_ : Union[str, Any] = '''1. Want to contribute a new model?''' with open(__snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ : List[str] = f.readlines() # Find the start of the list. lowercase_ : Optional[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase_ : str = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowercase_ : str = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(__snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__snake_case ) def lowercase ( ): with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowercase_ : List[Any] = f.read() lowercase_ : List[str] = REPLACE_PATTERNS['''init'''][0].search(__snake_case ).groups()[0] return packaging.version.parse(__snake_case ) def lowercase ( __snake_case : Optional[Any]=False ): lowercase_ : str = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowercase_ : Optional[Any] = default_version.base_version elif patch: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: lowercase_ : Optional[int] = F'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. lowercase_ : int = input(F'''Which version are you releasing? [{default_version}]''' ) if len(__snake_case ) == 0: lowercase_ : Dict = default_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case , patch=__snake_case ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def lowercase ( ): lowercase_ : List[Any] = get_version() lowercase_ : List[str] = F'''{current_version.major}.{current_version.minor + 1}.0.dev0''' lowercase_ : Any = current_version.base_version # Check with the user we got that right. lowercase_ : Tuple = input(F'''Which version are we developing now? [{dev_version}]''' ) if len(__snake_case ) == 0: lowercase_ : str = dev_version print(F'''Updating version to {version}.''' ) global_version_update(__snake_case ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": __A : int = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __A : Any = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
33
1
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) __A : str = logging.get_logger(__name__) # pylint: disable=invalid-name __A : int = ''' Examples: ```py >>> import torch >>> import numpy as np >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline >>> from transformers import pipeline >>> from diffusers.utils import load_image >>> def make_hint(image, depth_estimator): ... image = depth_estimator(image)["depth"] ... image = np.array(image) ... image = image[:, :, None] ... image = np.concatenate([image, image, image], axis=2) ... detected_map = torch.from_numpy(image).float() / 255.0 ... hint = detected_map.permute(2, 0, 1) ... return hint >>> depth_estimator = pipeline("depth-estimation") >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior = pipe_prior.to("cuda") >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16 ... ) >>> pipe = pipe.to("cuda") >>> img = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/cat.png" ... ).resize((768, 768)) >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda") >>> prompt = "A robot, 4k photo" >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature" >>> generator = torch.Generator(device="cuda").manual_seed(43) >>> image_emb, zero_image_emb = pipe_prior( ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator ... ).to_tuple() >>> images = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... hint=hint, ... num_inference_steps=50, ... generator=generator, ... height=768, ... width=768, ... ).images >>> images[0].save("robot_cat.png") ``` ''' def lowercase ( __snake_case : Any , __snake_case : int , __snake_case : Dict=8 ): lowercase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase_ : int = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class _UpperCAmelCase ( _A ): def __init__( self : List[Any] , A : UNetaDConditionModel , A : DDPMScheduler , A : VQModel , ) -> Any: super().__init__() self.register_modules( unet=A , scheduler=A , movq=A , ) lowercase_ : List[str] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def A ( self : Optional[Any] , A : Optional[Any] , A : Any , A : List[str] , A : Optional[int] , A : Optional[int] , A : List[str] ) -> List[str]: if latents is None: lowercase_ : str = randn_tensor(A , generator=A , device=A , dtype=A ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase_ : List[Any] = latents.to(A ) lowercase_ : Union[str, Any] = latents * scheduler.init_noise_sigma return latents def A ( self : Optional[Any] , A : Union[str, Any]=0 ) -> int: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) lowercase_ : str = torch.device(F'''cuda:{gpu_id}''' ) lowercase_ : List[Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A , A ) def A ( self : Union[str, Any] , A : Optional[int]=0 ) -> int: if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) lowercase_ : List[Any] = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase_ : Union[str, Any] = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase_ , lowercase_ : Any = cpu_offload_with_hook(A , A , prev_module_hook=A ) # We'll offload the last model manually. lowercase_ : Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def A ( self : str ) -> str: if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(A , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A ) def __call__( self : List[Any] , A : Union[torch.FloatTensor, List[torch.FloatTensor]] , A : Union[torch.FloatTensor, List[torch.FloatTensor]] , A : torch.FloatTensor , A : int = 5_12 , A : int = 5_12 , A : int = 1_00 , A : float = 4.0 , A : int = 1 , A : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A : Optional[torch.FloatTensor] = None , A : Optional[str] = "pil" , A : bool = True , ) -> List[Any]: lowercase_ : List[str] = self._execution_device lowercase_ : Optional[Any] = guidance_scale > 1.0 if isinstance(A , A ): lowercase_ : str = torch.cat(A , dim=0 ) if isinstance(A , A ): lowercase_ : Optional[int] = torch.cat(A , dim=0 ) if isinstance(A , A ): lowercase_ : Optional[Any] = torch.cat(A , dim=0 ) lowercase_ : List[str] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: lowercase_ : Any = image_embeds.repeat_interleave(A , dim=0 ) lowercase_ : Tuple = negative_image_embeds.repeat_interleave(A , dim=0 ) lowercase_ : Optional[int] = hint.repeat_interleave(A , dim=0 ) lowercase_ : Dict = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A ) lowercase_ : List[Any] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=A ) self.scheduler.set_timesteps(A , device=A ) lowercase_ : Optional[Any] = self.scheduler.timesteps lowercase_ : Union[str, Any] = self.movq.config.latent_channels lowercase_ , lowercase_ : Any = downscale_height_and_width(A , A , self.movq_scale_factor ) # create initial latent lowercase_ : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A , A , A , self.scheduler , ) for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance lowercase_ : Union[str, Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase_ : Dict = {'''image_embeds''': image_embeds, '''hint''': hint} lowercase_ : Dict = self.unet( sample=A , timestep=A , encoder_hidden_states=A , added_cond_kwargs=A , return_dict=A , )[0] if do_classifier_free_guidance: lowercase_ , lowercase_ : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) lowercase_ , lowercase_ : Optional[int] = noise_pred.chunk(2 ) lowercase_ , lowercase_ : List[str] = variance_pred.chunk(2 ) lowercase_ : Any = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase_ : Optional[int] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase_ , lowercase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase_ : Union[str, Any] = self.scheduler.step( A , A , A , generator=A , )[0] # post-processing lowercase_ : List[Any] = self.movq.decode(A , force_not_quantize=A )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase_ : List[Any] = image * 0.5 + 0.5 lowercase_ : Any = image.clamp(0 , 1 ) lowercase_ : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase_ : int = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
33
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def lowercase ( __snake_case : str , __snake_case : str , __snake_case : Optional[str] = None ): if version.parse(hfh.__version__ ).release < version.parse('''0.11.0''' ).release: # old versions of hfh don't url-encode the file path lowercase_ : Union[str, Any] = quote(__snake_case ) return hfh.hf_hub_url(__snake_case , __snake_case , repo_type='''dataset''' , revision=__snake_case )
33
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __A : Tuple = { '''configuration_convbert''': ['''CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ConvBertConfig''', '''ConvBertOnnxConfig'''], '''tokenization_convbert''': ['''ConvBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = ['''ConvBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ '''CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ConvBertForMaskedLM''', '''ConvBertForMultipleChoice''', '''ConvBertForQuestionAnswering''', '''ConvBertForSequenceClassification''', '''ConvBertForTokenClassification''', '''ConvBertLayer''', '''ConvBertModel''', '''ConvBertPreTrainedModel''', '''load_tf_weights_in_convbert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = [ '''TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFConvBertForMaskedLM''', '''TFConvBertForMultipleChoice''', '''TFConvBertForQuestionAnswering''', '''TFConvBertForSequenceClassification''', '''TFConvBertForTokenClassification''', '''TFConvBertLayer''', '''TFConvBertModel''', '''TFConvBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __A : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self : int , A : Tuple , A : int=3 , A : List[str]=32 , A : Dict=3 , A : Any=10 , A : Dict=[10, 20, 30, 40] , A : Optional[Any]=[1, 1, 2, 1] , A : Union[str, Any]=True , A : Optional[Any]=True , A : Any="relu" , A : Optional[Any]=3 , A : Tuple=None , ) -> Dict: lowercase_ : str = parent lowercase_ : List[Any] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : int = num_channels lowercase_ : int = embeddings_size lowercase_ : str = hidden_sizes lowercase_ : List[str] = depths lowercase_ : Dict = is_training lowercase_ : int = use_labels lowercase_ : Any = hidden_act lowercase_ : List[Any] = num_labels lowercase_ : Tuple = scope lowercase_ : Optional[Any] = len(A ) def A ( self : str ) -> Tuple: lowercase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : Union[str, Any] = None if self.use_labels: lowercase_ : List[str] = ids_tensor([self.batch_size] , self.num_labels ) lowercase_ : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Dict ) -> int: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def A ( self : str , A : Tuple , A : str , A : str ) -> str: lowercase_ : str = TFResNetModel(config=A ) lowercase_ : Union[str, Any] = model(A ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Any , A : int , A : List[Any] , A : Optional[Any] ) -> Optional[Any]: lowercase_ : Tuple = self.num_labels lowercase_ : Union[str, Any] = TFResNetForImageClassification(A ) lowercase_ : Tuple = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Union[str, Any] ) -> Tuple: lowercase_ : Tuple = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ : Dict = config_and_inputs lowercase_ : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _A , _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : List[Any] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : Any = False def A ( self : Union[str, Any] ) -> List[Any]: lowercase_ : int = TFResNetModelTester(self ) lowercase_ : str = ConfigTester(self , config_class=A , has_text_modality=A ) def A ( self : Dict ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Dict ) -> List[Any]: return @unittest.skip(reason='''ResNet does not use inputs_embeds''' ) def A ( self : Any ) -> Any: pass @unittest.skip(reason='''ResNet does not support input and output embeddings''' ) def A ( self : List[str] ) -> Optional[Any]: pass def A ( self : str ) -> Tuple: lowercase_ , lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(A ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : str = [*signature.parameters.keys()] lowercase_ : str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def A ( self : List[str] ) -> Tuple: lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def A ( self : List[Any] ) -> List[str]: def check_hidden_states_output(A : Union[str, Any] , A : int , A : List[Any] ): lowercase_ : int = model_class(A ) lowercase_ : Optional[Any] = model(**self._prepare_for_class(A , A ) ) lowercase_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase_ : Any = self.model_tester.num_stages self.assertEqual(len(A ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase_ , lowercase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase_ : List[str] = layer_type lowercase_ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ : Optional[Any] = True check_hidden_states_output(A , A , A ) def A ( self : Optional[int] ) -> Tuple: lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def A ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : Tuple = TFResNetModel.from_pretrained(A ) self.assertIsNotNone(A ) def lowercase ( ): lowercase_ : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase ): @cached_property def A ( self : Any ) -> Optional[int]: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A ( self : Any ) -> Optional[int]: lowercase_ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowercase_ : List[Any] = self.default_image_processor lowercase_ : Dict = prepare_img() lowercase_ : List[str] = image_processor(images=A , return_tensors='''tf''' ) # forward pass lowercase_ : Tuple = model(**A ) # verify the logits lowercase_ : Optional[int] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A ) lowercase_ : Optional[Any] = tf.constant([-11.1069, -9.7877, -8.3777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A , atol=1e-4 ) )
33
1
"""simple docstring""" from typing import Tuple, Union from ...modeling_outputs import BackboneOutput from ...modeling_utils import PreTrainedModel from ...utils import is_timm_available, is_torch_available, requires_backends from ...utils.backbone_utils import BackboneMixin from .configuration_timm_backbone import TimmBackboneConfig if is_timm_available(): import timm if is_torch_available(): from torch import Tensor class _UpperCAmelCase ( _A , _A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : Optional[Any] = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = TimmBackboneConfig def __init__( self : str , A : Any , **A : Union[str, Any] ) -> List[Any]: requires_backends(self , '''timm''' ) super().__init__(A ) lowercase_ : str = config if config.backbone is None: raise ValueError('''backbone is not set in the config. Please set it to a timm model name.''' ) if config.backbone not in timm.list_models(): raise ValueError(F'''backbone {config.backbone} is not supported by timm.''' ) if hasattr(A , '''out_features''' ) and config.out_features is not None: raise ValueError('''out_features is not supported by TimmBackbone. Please use out_indices instead.''' ) lowercase_ : Optional[Any] = getattr(A , '''use_pretrained_backbone''' , A ) if pretrained is None: raise ValueError('''use_pretrained_backbone is not set in the config. Please set it to True or False.''' ) # We just take the final layer by default. This matches the default for the transformers models. lowercase_ : Union[str, Any] = config.out_indices if getattr(A , '''out_indices''' , A ) is not None else (-1,) lowercase_ : Dict = timm.create_model( config.backbone , pretrained=A , features_only=config.features_only , in_chans=config.num_channels , out_indices=A , **A , ) # These are used to control the output of the model when called. If output_hidden_states is True, then # return_layers is modified to include all layers. lowercase_ : Dict = self._backbone.return_layers lowercase_ : Union[str, Any] = {layer['''module''']: str(A ) for i, layer in enumerate(self._backbone.feature_info.info )} super()._init_backbone(A ) @classmethod def A ( cls : Dict , A : Union[str, Any] , *A : Tuple , **A : str ) -> Optional[Any]: requires_backends(cls , ['''vision''', '''timm'''] ) from ...models.timm_backbone import TimmBackboneConfig lowercase_ : Optional[Any] = kwargs.pop('''config''' , TimmBackboneConfig() ) lowercase_ : Tuple = kwargs.pop('''use_timm_backbone''' , A ) if not use_timm: raise ValueError('''use_timm_backbone must be True for timm backbones''' ) lowercase_ : Any = kwargs.pop('''num_channels''' , config.num_channels ) lowercase_ : Any = kwargs.pop('''features_only''' , config.features_only ) lowercase_ : Any = kwargs.pop('''use_pretrained_backbone''' , config.use_pretrained_backbone ) lowercase_ : List[str] = kwargs.pop('''out_indices''' , config.out_indices ) lowercase_ : List[Any] = TimmBackboneConfig( backbone=A , num_channels=A , features_only=A , use_pretrained_backbone=A , out_indices=A , ) return super()._from_config(A , **A ) def A ( self : Tuple , A : int ) -> int: pass def A ( self : Dict , A : Union[str, Any] , A : int=None , A : Optional[int]=None , A : Union[str, Any]=None , **A : Union[str, Any] ) -> Union[BackboneOutput, Tuple[Tensor, ...]]: lowercase_ : str = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : List[Any] = output_attentions if output_attentions is not None else self.config.output_attentions if output_attentions: raise ValueError('''Cannot output attentions for timm backbones at the moment''' ) if output_hidden_states: # We modify the return layers to include all the stages of the backbone lowercase_ : Any = self._all_layers lowercase_ : Optional[int] = self._backbone(A , **A ) lowercase_ : Optional[Any] = self._return_layers lowercase_ : Dict = tuple(hidden_states[i] for i in self.out_indices ) else: lowercase_ : int = self._backbone(A , **A ) lowercase_ : str = None lowercase_ : Tuple = tuple(A ) lowercase_ : List[str] = tuple(A ) if hidden_states is not None else None if not return_dict: lowercase_ : Optional[int] = (feature_maps,) if output_hidden_states: lowercase_ : List[Any] = output + (hidden_states,) return output return BackboneOutput(feature_maps=A , hidden_states=A , attentions=A )
33
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin __A : Dict = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class _UpperCAmelCase ( unittest.TestCase , _A ): def A ( self : List[Any] ) -> Dict: lowercase_ : Optional[int] = load_tool('''text-question-answering''' ) self.tool.setup() lowercase_ : Union[str, Any] = load_tool('''text-question-answering''' , remote=A ) def A ( self : Any ) -> List[str]: lowercase_ : Union[str, Any] = self.tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : str ) -> List[str]: lowercase_ : int = self.remote_tool(A , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[Any] ) -> int: lowercase_ : Optional[Any] = self.tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' ) def A ( self : List[str] ) -> Optional[int]: lowercase_ : int = self.remote_tool(text=A , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(A , '''launched the BigScience Research Workshop''' )
33
1
"""simple docstring""" from __future__ import annotations def lowercase ( __snake_case : list , __snake_case : int | None = None , __snake_case : int | None = None ): if start is None: lowercase_ : Optional[Any] = 0 if end is None: lowercase_ : List[Any] = len(__snake_case ) - 1 if start >= end: return lowercase_ : str = (start + end) // 2 slowsort(__snake_case , __snake_case , __snake_case ) slowsort(__snake_case , mid + 1 , __snake_case ) if sequence[end] < sequence[mid]: lowercase_ , lowercase_ : Optional[int] = sequence[mid], sequence[end] slowsort(__snake_case , __snake_case , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
33
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _UpperCAmelCase ( TensorFormatter[Mapping, "torch.Tensor", Mapping] ): def __init__( self : Any , A : int=None , **A : str ) -> Union[str, Any]: super().__init__(features=A ) lowercase_ : Union[str, Any] = torch_tensor_kwargs import torch # noqa import torch at initialization def A ( self : Dict , A : int ) -> List[Any]: import torch if isinstance(A , A ) and column: if all( isinstance(A , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(A ) return column def A ( self : int , A : Any ) -> Optional[Any]: import torch if isinstance(A , (str, bytes, type(A )) ): return value elif isinstance(A , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() lowercase_ : Any = {} if isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): lowercase_ : Any = {'''dtype''': torch.intaa} elif isinstance(A , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): lowercase_ : Dict = {'''dtype''': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A , PIL.Image.Image ): lowercase_ : Dict = np.asarray(A ) return torch.tensor(A , **{**default_dtype, **self.torch_tensor_kwargs} ) def A ( self : Union[str, Any] , A : Optional[int] ) -> str: import torch # support for torch, tf, jax etc. if hasattr(A , '''__array__''' ) and not isinstance(A , torch.Tensor ): lowercase_ : Optional[int] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) elif isinstance(A , (list, tuple) ): return self._consolidate([self.recursive_tensorize(A ) for substruct in data_struct] ) return self._tensorize(A ) def A ( self : Dict , A : dict ) -> Tuple: return map_nested(self._recursive_tensorize , A , map_list=A ) def A ( self : str , A : pa.Table ) -> Mapping: lowercase_ : Optional[Any] = self.numpy_arrow_extractor().extract_row(A ) lowercase_ : str = self.python_features_decoder.decode_row(A ) return self.recursive_tensorize(A ) def A ( self : List[Any] , A : pa.Table ) -> "torch.Tensor": lowercase_ : List[str] = self.numpy_arrow_extractor().extract_column(A ) lowercase_ : str = self.python_features_decoder.decode_column(A , pa_table.column_names[0] ) lowercase_ : Optional[int] = self.recursive_tensorize(A ) lowercase_ : Any = self._consolidate(A ) return column def A ( self : List[str] , A : pa.Table ) -> Mapping: lowercase_ : Optional[int] = self.numpy_arrow_extractor().extract_batch(A ) lowercase_ : int = self.python_features_decoder.decode_batch(A ) lowercase_ : Dict = self.recursive_tensorize(A ) for column_name in batch: lowercase_ : Optional[Any] = self._consolidate(batch[column_name] ) return batch
33
1
"""simple docstring""" from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig __A : Any = logging.get_logger(__name__) # General docstring __A : int = '''RegNetConfig''' # Base docstring __A : Dict = '''facebook/regnet-y-040''' __A : Union[str, Any] = [1, 1_088, 7, 7] # Image classification docstring __A : List[str] = '''facebook/regnet-y-040''' __A : Optional[Any] = '''tabby, tabby cat''' __A : int = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : Tuple , A : int , A : int = 3 , A : int = 1 , A : int = 1 , A : Optional[str] = "relu" , **A : str , ) -> Dict: super().__init__(**A ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb lowercase_ : int = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) lowercase_ : Any = tf.keras.layers.ConvaD( filters=A , kernel_size=A , strides=A , padding='''VALID''' , groups=A , use_bias=A , name='''convolution''' , ) lowercase_ : Tuple = tf.keras.layers.BatchNormalization(epsilon=1e-5 , momentum=0.9 , name='''normalization''' ) lowercase_ : Union[str, Any] = ACTaFN[activation] if activation is not None else tf.identity def A ( self : Union[str, Any] , A : List[Any] ) -> Any: lowercase_ : Union[str, Any] = self.convolution(self.padding(A ) ) lowercase_ : Tuple = self.normalization(A ) lowercase_ : List[str] = self.activation(A ) return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : Union[str, Any] , A : RegNetConfig , **A : Dict ) -> Optional[Any]: super().__init__(**A ) lowercase_ : Any = config.num_channels lowercase_ : List[str] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='''embedder''' , ) def A ( self : Optional[int] , A : Dict ) -> Optional[Any]: lowercase_ : Dict = shape_list(A )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) lowercase_ : List[str] = tf.transpose(A , perm=(0, 2, 3, 1) ) lowercase_ : Union[str, Any] = self.embedder(A ) return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : Tuple , A : int , A : int = 2 , **A : Optional[Any] ) -> Tuple: super().__init__(**A ) lowercase_ : Tuple = tf.keras.layers.ConvaD( filters=A , kernel_size=1 , strides=A , use_bias=A , name='''convolution''' ) lowercase_ : Tuple = tf.keras.layers.BatchNormalization(epsilon=1e-5 , momentum=0.9 , name='''normalization''' ) def A ( self : Optional[Any] , A : tf.Tensor , A : bool = False ) -> tf.Tensor: return self.normalization(self.convolution(A ) , training=A ) class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : Union[str, Any] , A : int , A : int , **A : Union[str, Any] ) -> Optional[Any]: super().__init__(**A ) lowercase_ : Optional[Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=A , name='''pooler''' ) lowercase_ : Dict = [ tf.keras.layers.ConvaD(filters=A , kernel_size=1 , activation='''relu''' , name='''attention.0''' ), tf.keras.layers.ConvaD(filters=A , kernel_size=1 , activation='''sigmoid''' , name='''attention.2''' ), ] def A ( self : Tuple , A : Union[str, Any] ) -> List[Any]: # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] lowercase_ : Dict = self.pooler(A ) for layer_module in self.attention: lowercase_ : List[Any] = layer_module(A ) lowercase_ : List[str] = hidden_state * pooled return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : List[Any] , A : RegNetConfig , A : int , A : int , A : int = 1 , **A : Optional[Any] ) -> Dict: super().__init__(**A ) lowercase_ : str = in_channels != out_channels or stride != 1 lowercase_ : List[str] = max(1 , out_channels // config.groups_width ) lowercase_ : Dict = ( TFRegNetShortCut(A , stride=A , name='''shortcut''' ) if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. lowercase_ : Optional[Any] = [ TFRegNetConvLayer(A , kernel_size=1 , activation=config.hidden_act , name='''layer.0''' ), TFRegNetConvLayer( A , stride=A , groups=A , activation=config.hidden_act , name='''layer.1''' ), TFRegNetConvLayer(A , kernel_size=1 , activation=A , name='''layer.2''' ), ] lowercase_ : Union[str, Any] = ACTaFN[config.hidden_act] def A ( self : Tuple , A : List[Any] ) -> Any: lowercase_ : str = hidden_state for layer_module in self.layers: lowercase_ : Union[str, Any] = layer_module(A ) lowercase_ : List[Any] = self.shortcut(A ) hidden_state += residual lowercase_ : Tuple = self.activation(A ) return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : List[str] , A : RegNetConfig , A : int , A : int , A : int = 1 , **A : List[Any] ) -> str: super().__init__(**A ) lowercase_ : List[str] = in_channels != out_channels or stride != 1 lowercase_ : List[Any] = max(1 , out_channels // config.groups_width ) lowercase_ : Union[str, Any] = ( TFRegNetShortCut(A , stride=A , name='''shortcut''' ) if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''' ) ) lowercase_ : Optional[int] = [ TFRegNetConvLayer(A , kernel_size=1 , activation=config.hidden_act , name='''layer.0''' ), TFRegNetConvLayer( A , stride=A , groups=A , activation=config.hidden_act , name='''layer.1''' ), TFRegNetSELayer(A , reduced_channels=int(round(in_channels / 4 ) ) , name='''layer.2''' ), TFRegNetConvLayer(A , kernel_size=1 , activation=A , name='''layer.3''' ), ] lowercase_ : int = ACTaFN[config.hidden_act] def A ( self : int , A : Any ) -> Tuple: lowercase_ : Optional[int] = hidden_state for layer_module in self.layers: lowercase_ : Optional[Any] = layer_module(A ) lowercase_ : Optional[Any] = self.shortcut(A ) hidden_state += residual lowercase_ : Optional[Any] = self.activation(A ) return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : List[str] , A : RegNetConfig , A : int , A : int , A : int = 2 , A : int = 2 , **A : Optional[int] ) -> Tuple: super().__init__(**A ) lowercase_ : Dict = TFRegNetXLayer if config.layer_type == '''x''' else TFRegNetYLayer lowercase_ : Any = [ # downsampling is done in the first layer with stride of 2 layer(A , A , A , stride=A , name='''layers.0''' ), *[layer(A , A , A , name=F'''layers.{i+1}''' ) for i in range(depth - 1 )], ] def A ( self : List[Any] , A : Dict ) -> Any: for layer_module in self.layers: lowercase_ : Dict = layer_module(A ) return hidden_state class _UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self : int , A : RegNetConfig , **A : str ) -> List[str]: super().__init__(**A ) lowercase_ : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( A , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='''stages.0''' , ) ) lowercase_ : Tuple = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(A , config.depths[1:] ) ): self.stages.append(TFRegNetStage(A , A , A , depth=A , name=F'''stages.{i+1}''' ) ) def A ( self : Dict , A : tf.Tensor , A : bool = False , A : bool = True ) -> TFBaseModelOutputWithNoAttention: lowercase_ : Dict = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowercase_ : Dict = hidden_states + (hidden_state,) lowercase_ : int = stage_module(A ) if output_hidden_states: lowercase_ : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=A , hidden_states=A ) @keras_serializable class _UpperCAmelCase ( tf.keras.layers.Layer ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = RegNetConfig def __init__( self : Dict , A : int , **A : Dict ) -> Optional[Any]: super().__init__(**A ) lowercase_ : Optional[Any] = config lowercase_ : Any = TFRegNetEmbeddings(A , name='''embedder''' ) lowercase_ : str = TFRegNetEncoder(A , name='''encoder''' ) lowercase_ : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=A , name='''pooler''' ) @unpack_inputs def A ( self : Union[str, Any] , A : tf.Tensor , A : Optional[bool] = None , A : Optional[bool] = None , A : bool = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Dict = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : Optional[int] = self.embedder(A , training=A ) lowercase_ : int = self.encoder( A , output_hidden_states=A , return_dict=A , training=A ) lowercase_ : Union[str, Any] = encoder_outputs[0] lowercase_ : Optional[int] = self.pooler(A ) # Change to NCHW output format have uniformity in the modules lowercase_ : List[Any] = tf.transpose(A , perm=(0, 3, 1, 2) ) lowercase_ : Union[str, Any] = tf.transpose(A , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: lowercase_ : Any = tuple([tf.transpose(A , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = RegNetConfig SCREAMING_SNAKE_CASE_ : List[Any] = "regnet" SCREAMING_SNAKE_CASE_ : str = "pixel_values" @property def A ( self : Tuple ) -> List[Any]: return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} __A : Optional[Any] = R''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : int = R''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : Tuple , A : RegNetConfig , *A : Union[str, Any] , **A : List[str] ) -> Optional[Any]: super().__init__(A , *A , **A ) lowercase_ : Any = TFRegNetMainLayer(A , name='''regnet''' ) @unpack_inputs @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : Tuple , A : tf.Tensor , A : Optional[bool] = None , A : Optional[bool] = None , A : Dict=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: lowercase_ : List[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Dict = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[str] = self.regnet( pixel_values=A , output_hidden_states=A , return_dict=A , training=A , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A , _A ): def __init__( self : Union[str, Any] , A : RegNetConfig , *A : Optional[Any] , **A : Union[str, Any] ) -> int: super().__init__(A , *A , **A ) lowercase_ : Optional[Any] = config.num_labels lowercase_ : List[str] = TFRegNetMainLayer(A , name='''regnet''' ) # classification head lowercase_ : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='''classifier.1''' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : tf.Tensor = None , A : tf.Tensor = None , A : bool = None , A : bool = None , A : Union[str, Any]=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: lowercase_ : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : List[str] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : Any = self.regnet( A , output_hidden_states=A , return_dict=A , training=A ) lowercase_ : List[Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Any = self.classifier[0](A ) lowercase_ : Dict = self.classifier[1](A ) lowercase_ : Union[str, Any] = None if labels is None else self.hf_compute_loss(labels=A , logits=A ) if not return_dict: lowercase_ : Any = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=A , logits=A , hidden_states=outputs.hidden_states )
33
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
33
1
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_A ) class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True} ) SCREAMING_SNAKE_CASE_ : ClassVar[Features] = Features({"text": Value("string" )} ) SCREAMING_SNAKE_CASE_ : ClassVar[Features] = Features({} ) SCREAMING_SNAKE_CASE_ : str = "text" @property def A ( self : List[Any] ) -> Dict[str, str]: return {self.text_column: "text"}
33
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __A : Union[str, Any] = logging.get_logger(__name__) # General docstring __A : Tuple = '''MobileNetV1Config''' # Base docstring __A : Union[str, Any] = '''google/mobilenet_v1_1.0_224''' __A : Union[str, Any] = [1, 1_024, 7, 7] # Image classification docstring __A : Optional[Any] = '''google/mobilenet_v1_1.0_224''' __A : List[Any] = '''tabby, tabby cat''' __A : Union[str, Any] = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=None ): lowercase_ : str = {} if isinstance(__snake_case , __snake_case ): lowercase_ : Union[str, Any] = model.mobilenet_va else: lowercase_ : Optional[Any] = model lowercase_ : Union[str, Any] = '''MobilenetV1/Conv2d_0/''' lowercase_ : Union[str, Any] = backbone.conv_stem.convolution.weight lowercase_ : Optional[Any] = backbone.conv_stem.normalization.bias lowercase_ : Union[str, Any] = backbone.conv_stem.normalization.weight lowercase_ : Any = backbone.conv_stem.normalization.running_mean lowercase_ : int = backbone.conv_stem.normalization.running_var for i in range(1_3 ): lowercase_ : Optional[int] = i + 1 lowercase_ : Union[str, Any] = i * 2 lowercase_ : Optional[Any] = backbone.layer[pt_index] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' lowercase_ : str = pointer.convolution.weight lowercase_ : int = pointer.normalization.bias lowercase_ : Any = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Union[str, Any] = pointer.normalization.running_var lowercase_ : Any = backbone.layer[pt_index + 1] lowercase_ : Union[str, Any] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' lowercase_ : int = pointer.convolution.weight lowercase_ : str = pointer.normalization.bias lowercase_ : Tuple = pointer.normalization.weight lowercase_ : Dict = pointer.normalization.running_mean lowercase_ : Any = pointer.normalization.running_var if isinstance(__snake_case , __snake_case ): lowercase_ : Optional[Any] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase_ : Any = model.classifier.weight lowercase_ : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase_ : Tuple = tf.train.list_variables(__snake_case ) lowercase_ : int = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) lowercase_ : Optional[Any] = tf.train.load_variable(__snake_case , __snake_case ) lowercase_ : Optional[int] = array # Build TF to PyTorch weights loading map lowercase_ : Any = _build_tf_to_pytorch_map(__snake_case , __snake_case , __snake_case ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue lowercase_ : Union[str, Any] = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase_ : Any = np.transpose(__snake_case , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase_ : Optional[int] = array.squeeze().transpose() else: lowercase_ : Optional[int] = np.transpose(__snake_case , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) lowercase_ : str = torch.from_numpy(__snake_case ) tf_weights.pop(__snake_case , __snake_case ) tf_weights.pop(name + '''/RMSProp''' , __snake_case ) tf_weights.pop(name + '''/RMSProp_1''' , __snake_case ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __snake_case ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def lowercase ( __snake_case : torch.Tensor , __snake_case : nn.Convad ): lowercase_ , lowercase_ : Optional[int] = features.shape[-2:] lowercase_ , lowercase_ : str = conv_layer.stride lowercase_ , lowercase_ : Tuple = conv_layer.kernel_size if in_height % stride_height == 0: lowercase_ : Dict = max(kernel_height - stride_height , 0 ) else: lowercase_ : List[Any] = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase_ : str = max(kernel_width - stride_width , 0 ) else: lowercase_ : int = max(kernel_width - (in_width % stride_width) , 0 ) lowercase_ : int = pad_along_width // 2 lowercase_ : Union[str, Any] = pad_along_width - pad_left lowercase_ : Tuple = pad_along_height // 2 lowercase_ : List[str] = pad_along_height - pad_top lowercase_ : str = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__snake_case , __snake_case , '''constant''' , 0.0 ) class _UpperCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : MobileNetVaConfig , A : int , A : int , A : int , A : Optional[int] = 1 , A : Optional[int] = 1 , A : bool = False , A : Optional[bool] = True , A : Optional[bool or str] = True , ) -> None: super().__init__() lowercase_ : int = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) lowercase_ : Tuple = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase_ : int = nn.Convad( in_channels=A , out_channels=A , kernel_size=A , stride=A , padding=A , groups=A , bias=A , padding_mode='''zeros''' , ) if use_normalization: lowercase_ : Optional[Any] = nn.BatchNormad( num_features=A , eps=config.layer_norm_eps , momentum=0.9997 , affine=A , track_running_stats=A , ) else: lowercase_ : Union[str, Any] = None if use_activation: if isinstance(A , A ): lowercase_ : str = ACTaFN[use_activation] elif isinstance(config.hidden_act , A ): lowercase_ : Any = ACTaFN[config.hidden_act] else: lowercase_ : Tuple = config.hidden_act else: lowercase_ : Tuple = None def A ( self : str , A : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: lowercase_ : List[Any] = apply_tf_padding(A , self.convolution ) lowercase_ : Optional[int] = self.convolution(A ) if self.normalization is not None: lowercase_ : Union[str, Any] = self.normalization(A ) if self.activation is not None: lowercase_ : Optional[int] = self.activation(A ) return features class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Optional[int] = MobileNetVaConfig SCREAMING_SNAKE_CASE_ : int = load_tf_weights_in_mobilenet_va SCREAMING_SNAKE_CASE_ : Optional[Any] = "mobilenet_v1" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "pixel_values" SCREAMING_SNAKE_CASE_ : List[str] = False def A ( self : Any , A : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) __A : Union[str, Any] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __A : List[str] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : str , A : MobileNetVaConfig , A : bool = True ) -> int: super().__init__(A ) lowercase_ : Union[str, Any] = config lowercase_ : List[str] = 32 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase_ : Union[str, Any] = MobileNetVaConvLayer( A , in_channels=config.num_channels , out_channels=A , kernel_size=3 , stride=2 , ) lowercase_ : Optional[Any] = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase_ : List[Any] = nn.ModuleList() for i in range(13 ): lowercase_ : Dict = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase_ : str = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=3 , stride=strides[i] , groups=A , ) ) self.layer.append( MobileNetVaConvLayer( A , in_channels=A , out_channels=A , kernel_size=1 , ) ) lowercase_ : int = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def A ( self : Any , A : Optional[Any] ) -> Optional[int]: raise NotImplementedError @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : List[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: lowercase_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase_ : List[str] = self.conv_stem(A ) lowercase_ : Dict = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase_ : Optional[int] = layer_module(A ) if output_hidden_states: lowercase_ : str = all_hidden_states + (hidden_states,) lowercase_ : Tuple = hidden_states if self.pooler is not None: lowercase_ : Dict = torch.flatten(self.pooler(A ) , start_dim=1 ) else: lowercase_ : Optional[Any] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A , pooler_output=A , hidden_states=A , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , _A , ) class _UpperCAmelCase ( _A ): def __init__( self : List[str] , A : MobileNetVaConfig ) -> None: super().__init__(A ) lowercase_ : int = config.num_labels lowercase_ : List[str] = MobileNetVaModel(A ) lowercase_ : Union[str, Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase_ : Tuple = nn.Dropout(config.classifier_dropout_prob , inplace=A ) lowercase_ : int = nn.Linear(A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Optional[Any] , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , A : Optional[torch.Tensor] = None , A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: lowercase_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : List[Any] = self.mobilenet_va(A , output_hidden_states=A , return_dict=A ) lowercase_ : Union[str, Any] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Dict = self.classifier(self.dropout(A ) ) lowercase_ : int = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : Optional[Any] = '''single_label_classification''' else: lowercase_ : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : str = MSELoss() if self.num_labels == 1: lowercase_ : List[str] = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(A , A ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[Any] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : str = BCEWithLogitsLoss() lowercase_ : List[Any] = loss_fct(A , A ) if not return_dict: lowercase_ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A , logits=A , hidden_states=outputs.hidden_states , )
33
1
"""simple docstring""" __A : Union[str, Any] = [ (1_000, '''M'''), (900, '''CM'''), (500, '''D'''), (400, '''CD'''), (100, '''C'''), (90, '''XC'''), (50, '''L'''), (40, '''XL'''), (10, '''X'''), (9, '''IX'''), (5, '''V'''), (4, '''IV'''), (1, '''I'''), ] def lowercase ( __snake_case : str ): lowercase_ : List[Any] = {'''I''': 1, '''V''': 5, '''X''': 1_0, '''L''': 5_0, '''C''': 1_0_0, '''D''': 5_0_0, '''M''': 1_0_0_0} lowercase_ : Tuple = 0 lowercase_ : Optional[Any] = 0 while place < len(__snake_case ): if (place + 1 < len(__snake_case )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def lowercase ( __snake_case : int ): lowercase_ : List[Any] = [] for arabic, roman in ROMAN: ((lowercase_) , (lowercase_)) : Union[str, Any] = divmod(__snake_case , __snake_case ) result.append(roman * factor ) if number == 0: break return "".join(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" def lowercase ( __snake_case : list[int] ): lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): for j in range(i + 1 , __snake_case ): if numbers[j] < numbers[i]: lowercase_ , lowercase_ : Optional[int] = numbers[j], numbers[i] return numbers if __name__ == "__main__": __A : int = input('''Enter numbers separated by a comma:\n''').strip() __A : Any = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
33
1
"""simple docstring""" def lowercase ( __snake_case : List[str] , __snake_case : Optional[int] ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) lowercase_ : List[Any] = (boundary[1] - boundary[0]) / steps lowercase_ : List[str] = boundary[0] lowercase_ : int = boundary[1] lowercase_ : Optional[int] = make_points(__snake_case , __snake_case , __snake_case ) lowercase_ : str = 0.0 y += (h / 2.0) * f(__snake_case ) for i in x_i: # print(i) y += h * f(__snake_case ) y += (h / 2.0) * f(__snake_case ) return y def lowercase ( __snake_case : str , __snake_case : Tuple , __snake_case : Optional[int] ): lowercase_ : Dict = a + h while x < (b - h): yield x lowercase_ : Optional[int] = x + h def lowercase ( __snake_case : int ): # enter your function here lowercase_ : Any = (x - 0) * (x - 0) return y def lowercase ( ): lowercase_ : int = 0.0 # Lower bound of integration lowercase_ : List[Any] = 1.0 # Upper bound of integration lowercase_ : List[Any] = 10.0 # define number of steps or resolution lowercase_ : List[Any] = [a, b] # define boundary of integration lowercase_ : Optional[Any] = method_a(__snake_case , __snake_case ) print(F'''y = {y}''' ) if __name__ == "__main__": main()
33
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any] , A : Any , A : Tuple=7 , A : Tuple=3 , A : Optional[Any]=30 , A : List[Any]=4_00 , A : Tuple=True , A : Dict=None , A : List[str]=True , A : Optional[int]=[0.5, 0.5, 0.5] , A : Tuple=[0.5, 0.5, 0.5] , A : List[str]=True , A : List[Any]=1 / 2_55 , A : Union[str, Any]=True , ) -> Tuple: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase_ : Optional[int] = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} lowercase_ : Optional[int] = parent lowercase_ : str = batch_size lowercase_ : Tuple = num_channels lowercase_ : str = min_resolution lowercase_ : Any = max_resolution lowercase_ : str = do_resize lowercase_ : Any = size lowercase_ : Optional[int] = do_normalize lowercase_ : List[str] = image_mean lowercase_ : Optional[Any] = image_std lowercase_ : int = do_rescale lowercase_ : List[str] = rescale_factor lowercase_ : int = do_pad def A ( self : Any ) -> str: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Optional[Any] , A : int , A : int=False ) -> Tuple: if not batched: lowercase_ : Optional[int] = image_inputs[0] if isinstance(A , Image.Image ): lowercase_ , lowercase_ : int = image.size else: lowercase_ , lowercase_ : Tuple = image.shape[1], image.shape[2] if w < h: lowercase_ : int = int(self.size['''shortest_edge'''] * h / w ) lowercase_ : Optional[Any] = self.size['''shortest_edge'''] elif w > h: lowercase_ : Optional[Any] = self.size['''shortest_edge'''] lowercase_ : Optional[int] = int(self.size['''shortest_edge'''] * w / h ) else: lowercase_ : Any = self.size['''shortest_edge'''] lowercase_ : Any = self.size['''shortest_edge'''] else: lowercase_ : Tuple = [] for image in image_inputs: lowercase_ , lowercase_ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowercase_ : Union[str, Any] = max(A , key=lambda A : item[0] )[0] lowercase_ : Optional[Any] = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = YolosImageProcessor if is_vision_available() else None def A ( self : Optional[int] ) -> Optional[int]: lowercase_ : Optional[Any] = YolosImageProcessingTester(self ) @property def A ( self : str ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Optional[int] ) -> List[str]: lowercase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , '''image_mean''' ) ) self.assertTrue(hasattr(A , '''image_std''' ) ) self.assertTrue(hasattr(A , '''do_normalize''' ) ) self.assertTrue(hasattr(A , '''do_resize''' ) ) self.assertTrue(hasattr(A , '''size''' ) ) def A ( self : Dict ) -> Tuple: lowercase_ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowercase_ : Tuple = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , A ) def A ( self : Optional[int] ) -> Tuple: pass def A ( self : Tuple ) -> int: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowercase_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ , lowercase_ : Dict = self.image_processor_tester.get_expected_values(A , batched=A ) lowercase_ : str = image_processing(A , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : str ) -> Any: # Initialize image_processing lowercase_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowercase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : int = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Optional[int] = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[Any] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[int]: # Initialize image_processing lowercase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowercase_ : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase_ : Any = image_processing(A , return_tensors='''pt''' ).pixel_values lowercase_ , lowercase_ : List[str] = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Tuple ) -> Optional[Any]: # Initialize image_processings lowercase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) lowercase_ : Tuple = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowercase_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowercase_ : Union[str, Any] = image_processing_a.pad(A , return_tensors='''pt''' ) lowercase_ : List[Any] = image_processing_a(A , return_tensors='''pt''' ) self.assertTrue( torch.allclose(encoded_images_with_method['''pixel_values'''] , encoded_images['''pixel_values'''] , atol=1e-4 ) ) @slow def A ( self : str ) -> List[Any]: # prepare image and target lowercase_ : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowercase_ : List[Any] = json.loads(f.read() ) lowercase_ : Tuple = {'''image_id''': 3_97_69, '''annotations''': target} # encode them lowercase_ : Union[str, Any] = YolosImageProcessor.from_pretrained('''hustvl/yolos-small''' ) lowercase_ : List[Any] = image_processing(images=A , annotations=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Union[str, Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Union[str, Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : Tuple = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : List[str] = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : Any = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify orig_size lowercase_ : List[str] = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : Optional[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) ) @slow def A ( self : List[Any] ) -> Dict: # prepare image, target and masks_path lowercase_ : Tuple = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowercase_ : str = json.loads(f.read() ) lowercase_ : int = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} lowercase_ : List[Any] = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowercase_ : int = YolosImageProcessor(format='''coco_panoptic''' ) lowercase_ : Any = image_processing(images=A , annotations=A , masks_path=A , return_tensors='''pt''' ) # verify pixel values lowercase_ : Optional[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , A ) lowercase_ : Tuple = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowercase_ : List[Any] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , A ) ) # verify boxes lowercase_ : str = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , A ) lowercase_ : List[str] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , A , atol=1e-3 ) ) # verify image_id lowercase_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , A ) ) # verify is_crowd lowercase_ : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , A ) ) # verify class_labels lowercase_ : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , A ) ) # verify masks lowercase_ : Dict = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , A ) # verify orig_size lowercase_ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , A ) ) # verify size lowercase_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , A ) )
33
1
"""simple docstring""" import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 __A : Union[str, Any] = get_tests_dir('''fixtures''') class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Optional[Any] ) -> List[str]: # A mock response for an HTTP head request to emulate server down lowercase_ : Tuple = mock.Mock() lowercase_ : Tuple = 5_00 lowercase_ : Any = {} lowercase_ : Union[str, Any] = HTTPError lowercase_ : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowercase_ : Dict = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=A ) as mock_head: lowercase_ : Tuple = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ) -> int: # This test is for deprecated behavior and can be removed in v5 lowercase_ : Union[str, Any] = ViTImageProcessor.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json''' ) def A ( self : str ) -> List[str]: with self.assertRaises(A ): # config is in subfolder, the following should not work without specifying the subfolder lowercase_ : str = AutoImageProcessor.from_pretrained('''hf-internal-testing/stable-diffusion-all-variants''' ) lowercase_ : Optional[Any] = AutoImageProcessor.from_pretrained( '''hf-internal-testing/stable-diffusion-all-variants''' , subfolder='''feature_extractor''' ) self.assertIsNotNone(A ) @is_staging_test class _UpperCAmelCase ( unittest.TestCase ): @classmethod def A ( cls : Union[str, Any] ) -> int: lowercase_ : Dict = TOKEN HfFolder.save_token(A ) @classmethod def A ( cls : Any ) -> int: try: delete_repo(token=cls._token , repo_id='''test-image-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-image-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-image-processor''' ) except HTTPError: pass def A ( self : List[str] ) -> Dict: lowercase_ : List[Any] = ViTImageProcessor.from_pretrained(A ) image_processor.push_to_hub('''test-image-processor''' , use_auth_token=self._token ) lowercase_ : Union[str, Any] = ViTImageProcessor.from_pretrained(F'''{USER}/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( A , repo_id='''test-image-processor''' , push_to_hub=A , use_auth_token=self._token ) lowercase_ : str = ViTImageProcessor.from_pretrained(F'''{USER}/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) def A ( self : str ) -> List[Any]: lowercase_ : List[Any] = ViTImageProcessor.from_pretrained(A ) image_processor.push_to_hub('''valid_org/test-image-processor''' , use_auth_token=self._token ) lowercase_ : Union[str, Any] = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( A , repo_id='''valid_org/test-image-processor-org''' , push_to_hub=A , use_auth_token=self._token ) lowercase_ : str = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor-org''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) def A ( self : Union[str, Any] ) -> Tuple: CustomImageProcessor.register_for_auto_class() lowercase_ : List[str] = CustomImageProcessor.from_pretrained(A ) image_processor.push_to_hub('''test-dynamic-image-processor''' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {'''AutoImageProcessor''': '''custom_image_processing.CustomImageProcessor'''} , ) lowercase_ : Dict = AutoImageProcessor.from_pretrained( F'''{USER}/test-dynamic-image-processor''' , trust_remote_code=A ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , '''CustomImageProcessor''' )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0 ): lowercase_ : str = 0 lowercase_ : List[Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
33
1
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--txt2img_unclip''', default='''kakaobrain/karlo-v1-alpha''', type=str, required=False, help='''The pretrained txt2img unclip.''', ) __A : str = parser.parse_args() __A : List[Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __A : Dict = CLIPImageProcessor() __A : Union[str, Any] = CLIPVisionModelWithProjection.from_pretrained('''openai/clip-vit-large-patch14''') __A : List[str] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
33
1
"""simple docstring""" def lowercase ( __snake_case : list ): for i in range(len(__snake_case ) - 1 , 0 , -1 ): lowercase_ : int = False for j in range(__snake_case , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: lowercase_ , lowercase_ : Dict = unsorted[j - 1], unsorted[j] lowercase_ : List[str] = True for j in range(__snake_case ): if unsorted[j] > unsorted[j + 1]: lowercase_ , lowercase_ : int = unsorted[j + 1], unsorted[j] lowercase_ : Dict = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() __A : List[str] = input('''Enter numbers separated by a comma:\n''').strip() __A : str = [int(item) for item in user_input.split(''',''')] print(F"""{cocktail_shaker_sort(unsorted) = }""")
33
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Any = KandinskyVaaControlnetImgaImgPipeline SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : str = ["image_embeds", "negative_image_embeds", "image", "hint"] SCREAMING_SNAKE_CASE_ : Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE_ : Dict = False @property def A ( self : Any ) -> Any: return 32 @property def A ( self : Optional[int] ) -> Any: return 32 @property def A ( self : Dict ) -> int: return self.time_input_dim @property def A ( self : Tuple ) -> str: return self.time_input_dim * 4 @property def A ( self : Any ) -> str: return 1_00 @property def A ( self : str ) -> List[str]: torch.manual_seed(0 ) lowercase_ : List[Any] = { '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } lowercase_ : Dict = UNetaDConditionModel(**A ) return model @property def A ( self : Optional[Any] ) -> Union[str, Any]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A ( self : List[Any] ) -> Dict: torch.manual_seed(0 ) lowercase_ : int = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Tuple = self.dummy_unet lowercase_ : int = self.dummy_movq lowercase_ : List[Any] = { '''num_train_timesteps''': 10_00, '''beta_schedule''': '''linear''', '''beta_start''': 0.00085, '''beta_end''': 0.012, '''clip_sample''': False, '''set_alpha_to_one''': False, '''steps_offset''': 0, '''prediction_type''': '''epsilon''', '''thresholding''': False, } lowercase_ : str = DDIMScheduler(**A ) lowercase_ : Tuple = { '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A ( self : Optional[int] , A : int , A : List[str]=0 ) -> int: lowercase_ : str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A ) ).to(A ) lowercase_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A ) # create init_image lowercase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) lowercase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase_ : Optional[Any] = Image.fromarray(np.uinta(A ) ).convert('''RGB''' ).resize((2_56, 2_56) ) # create hint lowercase_ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A ) ).to(A ) if str(A ).startswith('''mps''' ): lowercase_ : Optional[Any] = torch.manual_seed(A ) else: lowercase_ : List[Any] = torch.Generator(device=A ).manual_seed(A ) lowercase_ : Dict = { '''image''': init_image, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 10, '''guidance_scale''': 7.0, '''strength''': 0.2, '''output_type''': '''np''', } return inputs def A ( self : Any ) -> List[Any]: lowercase_ : List[str] = '''cpu''' lowercase_ : Any = self.get_dummy_components() lowercase_ : Any = self.pipeline_class(**A ) lowercase_ : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) lowercase_ : Dict = pipe(**self.get_dummy_inputs(A ) ) lowercase_ : str = output.images lowercase_ : int = pipe( **self.get_dummy_inputs(A ) , return_dict=A , )[0] lowercase_ : Dict = image[0, -3:, -3:, -1] lowercase_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase_ : List[str] = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Tuple ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Any ) -> Optional[int]: lowercase_ : Dict = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy''' ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) lowercase_ : Optional[int] = init_image.resize((5_12, 5_12) ) lowercase_ : Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) lowercase_ : Optional[int] = torch.from_numpy(np.array(A ) ).float() / 255.0 lowercase_ : Tuple = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowercase_ : Optional[Any] = '''A robot, 4k photo''' lowercase_ : Tuple = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(A ) lowercase_ : Dict = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) lowercase_ : int = pipeline.to(A ) pipeline.set_progress_bar_config(disable=A ) lowercase_ : Tuple = torch.Generator(device='''cpu''' ).manual_seed(0 ) lowercase_ , lowercase_ : int = pipe_prior( A , image=A , strength=0.85 , generator=A , negative_prompt='''''' , ).to_tuple() lowercase_ : str = pipeline( image=A , image_embeds=A , negative_image_embeds=A , hint=A , generator=A , num_inference_steps=1_00 , height=5_12 , width=5_12 , strength=0.5 , output_type='''np''' , ) lowercase_ : Optional[Any] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(A , A )
33
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class _UpperCAmelCase ( unittest.TestCase ): def __init__( self : Dict , A : List[Any] , A : Tuple=7 , A : Dict=3 , A : Optional[Any]=18 , A : Union[str, Any]=30 , A : List[str]=4_00 , A : List[Any]=True , A : Union[str, Any]=32 , A : Any=True , ) -> List[Any]: lowercase_ : Optional[Any] = parent lowercase_ : Optional[int] = batch_size lowercase_ : Any = num_channels lowercase_ : List[str] = image_size lowercase_ : Optional[int] = min_resolution lowercase_ : Dict = max_resolution lowercase_ : str = do_resize lowercase_ : Optional[Any] = size_divisor lowercase_ : List[Any] = do_rescale def A ( self : Dict ) -> Optional[int]: return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class _UpperCAmelCase ( _A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : List[Any] = GLPNImageProcessor if is_vision_available() else None def A ( self : Union[str, Any] ) -> List[str]: lowercase_ : Any = GLPNImageProcessingTester(self ) @property def A ( self : int ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def A ( self : List[str] ) -> int: lowercase_ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , '''do_resize''' ) ) self.assertTrue(hasattr(A , '''size_divisor''' ) ) self.assertTrue(hasattr(A , '''resample''' ) ) self.assertTrue(hasattr(A , '''do_rescale''' ) ) def A ( self : Any ) -> str: pass def A ( self : Tuple ) -> Tuple: # Initialize image_processing lowercase_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase_ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase_ : str = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A ( self : List[Any] ) -> Tuple: # Initialize image_processing lowercase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A ( self : Optional[int] ) -> List[Any]: # Initialize image_processing lowercase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0_0 ): lowercase_ , lowercase_ : str = 1, 1 lowercase_ : List[str] = 2 while True: lowercase_ : Tuple = 0 lowercase_ : List[Any] = fa + fa lowercase_ , lowercase_ : Optional[int] = fa, f index += 1 for _ in str(__snake_case ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" import pytest import datasets # Import fixture modules as plugins __A : Dict = ['''tests.fixtures.files''', '''tests.fixtures.hub''', '''tests.fixtures.fsspec'''] def lowercase ( __snake_case : List[Any] , __snake_case : Dict ): # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def lowercase ( __snake_case : List[str] ): config.addinivalue_line('''markers''' , '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=__snake_case ) def lowercase ( __snake_case : List[Any] , __snake_case : Dict ): # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? lowercase_ : Optional[int] = tmp_path_factory.getbasetemp() / '''cache''' lowercase_ : List[str] = test_hf_cache_home / '''datasets''' lowercase_ : Any = test_hf_cache_home / '''metrics''' lowercase_ : Optional[int] = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''' , str(__snake_case ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''' , str(__snake_case ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''' , str(__snake_case ) ) lowercase_ : Optional[Any] = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''' , str(__snake_case ) ) lowercase_ : Tuple = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(__snake_case ) ) @pytest.fixture(autouse=__snake_case , scope='''session''' ) def lowercase ( ): datasets.disable_progress_bar() @pytest.fixture(autouse=__snake_case ) def lowercase ( __snake_case : List[Any] ): # don't take tests into account when counting downloads monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''' , __snake_case ) @pytest.fixture def lowercase ( __snake_case : Tuple ): # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''' , __snake_case )
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : Union[str, Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = "vit_mae" def __init__( self : Dict , A : List[str]=7_68 , A : Any=12 , A : Union[str, Any]=12 , A : Tuple=30_72 , A : Any="gelu" , A : Tuple=0.0 , A : List[str]=0.0 , A : Tuple=0.02 , A : Tuple=1e-12 , A : int=2_24 , A : Dict=16 , A : int=3 , A : Tuple=True , A : Tuple=16 , A : Optional[Any]=5_12 , A : Union[str, Any]=8 , A : List[Any]=20_48 , A : Dict=0.75 , A : Any=False , **A : Optional[int] , ) -> Union[str, Any]: super().__init__(**A ) lowercase_ : List[Any] = hidden_size lowercase_ : str = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Any = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : List[Any] = hidden_dropout_prob lowercase_ : int = attention_probs_dropout_prob lowercase_ : int = initializer_range lowercase_ : Dict = layer_norm_eps lowercase_ : Optional[Any] = image_size lowercase_ : str = patch_size lowercase_ : Dict = num_channels lowercase_ : Any = qkv_bias lowercase_ : Union[str, Any] = decoder_num_attention_heads lowercase_ : Optional[Any] = decoder_hidden_size lowercase_ : List[str] = decoder_num_hidden_layers lowercase_ : List[Any] = decoder_intermediate_size lowercase_ : Optional[Any] = mask_ratio lowercase_ : Optional[Any] = norm_pix_loss
33
1
"""simple docstring""" import operator as op __A : Union[str, Any] = '''scaler.pt''' __A : List[str] = '''pytorch_model''' __A : List[Any] = '''random_states''' __A : int = '''optimizer''' __A : List[Any] = '''scheduler''' __A : Tuple = '''pytorch_model.bin''' __A : Union[str, Any] = '''pytorch_model.bin.index.json''' __A : Any = '''model.safetensors''' __A : int = '''model.safetensors.index.json''' __A : Any = '''1.10.2''' __A : List[Any] = '''py38''' __A : Union[str, Any] = '''4.17.0''' __A : Optional[Any] = ['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] __A : int = ['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] __A : int = ['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] __A : Tuple = ['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] __A : Optional[Any] = ['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] __A : Optional[int] = '''2.0.1''' __A : Union[str, Any] = ['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] __A : Optional[Any] = ['''default''', '''reduce-overhead''', '''max-autotune'''] __A : Optional[Any] = {'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 __A : Dict = [ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] __A : List[str] = ['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] __A : Optional[Any] = ['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
33
"""simple docstring""" def lowercase ( __snake_case : int ): if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: lowercase_ : Dict = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase ( __snake_case : int ): lowercase_ : str = 0 lowercase_ : List[str] = 2 while digits < n: index += 1 lowercase_ : Any = len(str(fibonacci(__snake_case ) ) ) return index def lowercase ( __snake_case : int = 1_0_0_0 ): return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
33
1
"""simple docstring""" def lowercase ( __snake_case : Union[str, Any] , __snake_case : Tuple ): lowercase_ : Tuple = [1] for i in range(2 , __snake_case ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" lowercase_ : str = [] lowercase_ : str = list(range(__snake_case ) ) # Find permutation while factorials: lowercase_ : int = factorials.pop() lowercase_ , lowercase_ : List[Any] = divmod(__snake_case , __snake_case ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __A : List[str] = { '''configuration_mobilenet_v2''': [ '''MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileNetV2Config''', '''MobileNetV2OnnxConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Dict = ['''MobileNetV2FeatureExtractor'''] __A : Optional[int] = ['''MobileNetV2ImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileNetV2ForImageClassification''', '''MobileNetV2ForSemanticSegmentation''', '''MobileNetV2Model''', '''MobileNetV2PreTrainedModel''', '''load_tf_weights_in_mobilenet_v2''', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
1
"""simple docstring""" def lowercase ( __snake_case : str , __snake_case : str ): if not (isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case )): raise ValueError('''longest_common_substring() takes two strings for inputs''' ) lowercase_ : Dict = len(__snake_case ) lowercase_ : Optional[int] = len(__snake_case ) lowercase_ : Any = [[0] * (texta_length + 1) for _ in range(texta_length + 1 )] lowercase_ : Dict = 0 lowercase_ : str = 0 for i in range(1 , texta_length + 1 ): for j in range(1 , texta_length + 1 ): if texta[i - 1] == texta[j - 1]: lowercase_ : Optional[Any] = 1 + dp[i - 1][j - 1] if dp[i][j] > ans_length: lowercase_ : List[str] = i lowercase_ : Any = dp[i][j] return texta[ans_index - ans_length : ans_index] if __name__ == "__main__": import doctest doctest.testmod()
33
"""simple docstring""" from __future__ import annotations __A : List[Any] = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __A : str = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] lowercase_ : List[Any] = len(__snake_case ) for i in range(__snake_case ): lowercase_ : float = -1 for j in range(i + 1 , __snake_case ): if arr[i] < arr[j]: lowercase_ : List[str] = arr[j] break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = [] for i, outer in enumerate(__snake_case ): lowercase_ : float = -1 for inner in arr[i + 1 :]: if outer < inner: lowercase_ : List[Any] = inner break result.append(__snake_case ) return result def lowercase ( __snake_case : list[float] ): lowercase_ : List[str] = len(__snake_case ) lowercase_ : list[float] = [] lowercase_ : list[float] = [-1] * arr_size for index in reversed(range(__snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowercase_ : Optional[Any] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __A : int = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
33
1